var/home/core/zuul-output/0000755000175000017500000000000015066760236014540 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015066770706015507 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004652220315066770700017711 0ustar rootrootSep 30 13:56:10 crc systemd[1]: Starting Kubernetes Kubelet... Sep 30 13:56:10 crc restorecon[4678]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:10 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 13:56:11 crc restorecon[4678]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 13:56:11 crc restorecon[4678]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Sep 30 13:56:11 crc kubenswrapper[4840]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 30 13:56:11 crc kubenswrapper[4840]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Sep 30 13:56:11 crc kubenswrapper[4840]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 30 13:56:11 crc kubenswrapper[4840]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 30 13:56:11 crc kubenswrapper[4840]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Sep 30 13:56:11 crc kubenswrapper[4840]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.832175 4840 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.835789 4840 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.835810 4840 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.835814 4840 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.835819 4840 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.835825 4840 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.835831 4840 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.835836 4840 feature_gate.go:330] unrecognized feature gate: OVNObservability Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.835841 4840 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.835845 4840 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.835850 4840 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.835855 4840 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.835864 4840 feature_gate.go:330] unrecognized feature gate: Example Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.835868 4840 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.835871 4840 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.835875 4840 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.835879 4840 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.835882 4840 feature_gate.go:330] unrecognized feature gate: NewOLM Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.835886 4840 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.835890 4840 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.835893 4840 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.835897 4840 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.835901 4840 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.835905 4840 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.835909 4840 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.835916 4840 feature_gate.go:330] unrecognized feature gate: PinnedImages Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.835920 4840 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.835924 4840 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.835928 4840 feature_gate.go:330] unrecognized feature gate: SignatureStores Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.835931 4840 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.835935 4840 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.835940 4840 feature_gate.go:330] unrecognized feature gate: PlatformOperators Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.835945 4840 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.835948 4840 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.835953 4840 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.835958 4840 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.835961 4840 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.835965 4840 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.835972 4840 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.835976 4840 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.836038 4840 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.836045 4840 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.836063 4840 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.836067 4840 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.836072 4840 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.836077 4840 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.836083 4840 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.836087 4840 feature_gate.go:330] unrecognized feature gate: GatewayAPI Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.836091 4840 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.836095 4840 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.836103 4840 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.836107 4840 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.836111 4840 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.836145 4840 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.836149 4840 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.836152 4840 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.836158 4840 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.836163 4840 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.836167 4840 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.836171 4840 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.836176 4840 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.836180 4840 feature_gate.go:330] unrecognized feature gate: InsightsConfig Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.836184 4840 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.836188 4840 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.836193 4840 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.836232 4840 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.836263 4840 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.836271 4840 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.836471 4840 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.836480 4840 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.836492 4840 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.836499 4840 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836621 4840 flags.go:64] FLAG: --address="0.0.0.0" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836633 4840 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836649 4840 flags.go:64] FLAG: --anonymous-auth="true" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836655 4840 flags.go:64] FLAG: --application-metrics-count-limit="100" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836661 4840 flags.go:64] FLAG: --authentication-token-webhook="false" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836666 4840 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836672 4840 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836678 4840 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836683 4840 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836687 4840 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836692 4840 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836697 4840 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836701 4840 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836707 4840 flags.go:64] FLAG: --cgroup-root="" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836712 4840 flags.go:64] FLAG: --cgroups-per-qos="true" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836717 4840 flags.go:64] FLAG: --client-ca-file="" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836721 4840 flags.go:64] FLAG: --cloud-config="" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836726 4840 flags.go:64] FLAG: --cloud-provider="" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836731 4840 flags.go:64] FLAG: --cluster-dns="[]" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836737 4840 flags.go:64] FLAG: --cluster-domain="" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836741 4840 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836746 4840 flags.go:64] FLAG: --config-dir="" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836750 4840 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836755 4840 flags.go:64] FLAG: --container-log-max-files="5" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836760 4840 flags.go:64] FLAG: --container-log-max-size="10Mi" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836764 4840 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836768 4840 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836773 4840 flags.go:64] FLAG: --containerd-namespace="k8s.io" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836777 4840 flags.go:64] FLAG: --contention-profiling="false" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836782 4840 flags.go:64] FLAG: --cpu-cfs-quota="true" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836786 4840 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836793 4840 flags.go:64] FLAG: --cpu-manager-policy="none" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836797 4840 flags.go:64] FLAG: --cpu-manager-policy-options="" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836803 4840 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836808 4840 flags.go:64] FLAG: --enable-controller-attach-detach="true" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836812 4840 flags.go:64] FLAG: --enable-debugging-handlers="true" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836816 4840 flags.go:64] FLAG: --enable-load-reader="false" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836820 4840 flags.go:64] FLAG: --enable-server="true" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836825 4840 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836831 4840 flags.go:64] FLAG: --event-burst="100" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836835 4840 flags.go:64] FLAG: --event-qps="50" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836839 4840 flags.go:64] FLAG: --event-storage-age-limit="default=0" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836843 4840 flags.go:64] FLAG: --event-storage-event-limit="default=0" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836847 4840 flags.go:64] FLAG: --eviction-hard="" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836853 4840 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836857 4840 flags.go:64] FLAG: --eviction-minimum-reclaim="" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836862 4840 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836866 4840 flags.go:64] FLAG: --eviction-soft="" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836870 4840 flags.go:64] FLAG: --eviction-soft-grace-period="" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836874 4840 flags.go:64] FLAG: --exit-on-lock-contention="false" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836879 4840 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836883 4840 flags.go:64] FLAG: --experimental-mounter-path="" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836887 4840 flags.go:64] FLAG: --fail-cgroupv1="false" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836891 4840 flags.go:64] FLAG: --fail-swap-on="true" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836895 4840 flags.go:64] FLAG: --feature-gates="" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836900 4840 flags.go:64] FLAG: --file-check-frequency="20s" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836904 4840 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836908 4840 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836913 4840 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836917 4840 flags.go:64] FLAG: --healthz-port="10248" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836921 4840 flags.go:64] FLAG: --help="false" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836926 4840 flags.go:64] FLAG: --hostname-override="" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836930 4840 flags.go:64] FLAG: --housekeeping-interval="10s" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836935 4840 flags.go:64] FLAG: --http-check-frequency="20s" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836939 4840 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836943 4840 flags.go:64] FLAG: --image-credential-provider-config="" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836947 4840 flags.go:64] FLAG: --image-gc-high-threshold="85" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836951 4840 flags.go:64] FLAG: --image-gc-low-threshold="80" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836956 4840 flags.go:64] FLAG: --image-service-endpoint="" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836960 4840 flags.go:64] FLAG: --kernel-memcg-notification="false" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836964 4840 flags.go:64] FLAG: --kube-api-burst="100" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836968 4840 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836972 4840 flags.go:64] FLAG: --kube-api-qps="50" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836976 4840 flags.go:64] FLAG: --kube-reserved="" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836980 4840 flags.go:64] FLAG: --kube-reserved-cgroup="" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836984 4840 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836989 4840 flags.go:64] FLAG: --kubelet-cgroups="" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836993 4840 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.836998 4840 flags.go:64] FLAG: --lock-file="" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837016 4840 flags.go:64] FLAG: --log-cadvisor-usage="false" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837020 4840 flags.go:64] FLAG: --log-flush-frequency="5s" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837025 4840 flags.go:64] FLAG: --log-json-info-buffer-size="0" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837031 4840 flags.go:64] FLAG: --log-json-split-stream="false" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837036 4840 flags.go:64] FLAG: --log-text-info-buffer-size="0" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837041 4840 flags.go:64] FLAG: --log-text-split-stream="false" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837045 4840 flags.go:64] FLAG: --logging-format="text" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837049 4840 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837054 4840 flags.go:64] FLAG: --make-iptables-util-chains="true" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837058 4840 flags.go:64] FLAG: --manifest-url="" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837062 4840 flags.go:64] FLAG: --manifest-url-header="" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837068 4840 flags.go:64] FLAG: --max-housekeeping-interval="15s" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837072 4840 flags.go:64] FLAG: --max-open-files="1000000" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837077 4840 flags.go:64] FLAG: --max-pods="110" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837081 4840 flags.go:64] FLAG: --maximum-dead-containers="-1" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837086 4840 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837090 4840 flags.go:64] FLAG: --memory-manager-policy="None" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837094 4840 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837099 4840 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837103 4840 flags.go:64] FLAG: --node-ip="192.168.126.11" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837107 4840 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837121 4840 flags.go:64] FLAG: --node-status-max-images="50" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837125 4840 flags.go:64] FLAG: --node-status-update-frequency="10s" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837129 4840 flags.go:64] FLAG: --oom-score-adj="-999" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837133 4840 flags.go:64] FLAG: --pod-cidr="" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837138 4840 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837146 4840 flags.go:64] FLAG: --pod-manifest-path="" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837150 4840 flags.go:64] FLAG: --pod-max-pids="-1" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837154 4840 flags.go:64] FLAG: --pods-per-core="0" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837158 4840 flags.go:64] FLAG: --port="10250" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837163 4840 flags.go:64] FLAG: --protect-kernel-defaults="false" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837167 4840 flags.go:64] FLAG: --provider-id="" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837171 4840 flags.go:64] FLAG: --qos-reserved="" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837175 4840 flags.go:64] FLAG: --read-only-port="10255" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837179 4840 flags.go:64] FLAG: --register-node="true" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837183 4840 flags.go:64] FLAG: --register-schedulable="true" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837188 4840 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837195 4840 flags.go:64] FLAG: --registry-burst="10" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837199 4840 flags.go:64] FLAG: --registry-qps="5" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837203 4840 flags.go:64] FLAG: --reserved-cpus="" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837209 4840 flags.go:64] FLAG: --reserved-memory="" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837216 4840 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837220 4840 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837225 4840 flags.go:64] FLAG: --rotate-certificates="false" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837230 4840 flags.go:64] FLAG: --rotate-server-certificates="false" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837235 4840 flags.go:64] FLAG: --runonce="false" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837240 4840 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837244 4840 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837249 4840 flags.go:64] FLAG: --seccomp-default="false" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837253 4840 flags.go:64] FLAG: --serialize-image-pulls="true" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837257 4840 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837261 4840 flags.go:64] FLAG: --storage-driver-db="cadvisor" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837265 4840 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837270 4840 flags.go:64] FLAG: --storage-driver-password="root" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837274 4840 flags.go:64] FLAG: --storage-driver-secure="false" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837278 4840 flags.go:64] FLAG: --storage-driver-table="stats" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837282 4840 flags.go:64] FLAG: --storage-driver-user="root" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837286 4840 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837291 4840 flags.go:64] FLAG: --sync-frequency="1m0s" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837295 4840 flags.go:64] FLAG: --system-cgroups="" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837299 4840 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837307 4840 flags.go:64] FLAG: --system-reserved-cgroup="" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837311 4840 flags.go:64] FLAG: --tls-cert-file="" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837315 4840 flags.go:64] FLAG: --tls-cipher-suites="[]" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837321 4840 flags.go:64] FLAG: --tls-min-version="" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837325 4840 flags.go:64] FLAG: --tls-private-key-file="" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837330 4840 flags.go:64] FLAG: --topology-manager-policy="none" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837335 4840 flags.go:64] FLAG: --topology-manager-policy-options="" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837339 4840 flags.go:64] FLAG: --topology-manager-scope="container" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837344 4840 flags.go:64] FLAG: --v="2" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837350 4840 flags.go:64] FLAG: --version="false" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837356 4840 flags.go:64] FLAG: --vmodule="" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837361 4840 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837366 4840 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837473 4840 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837478 4840 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837483 4840 feature_gate.go:330] unrecognized feature gate: PlatformOperators Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837487 4840 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837491 4840 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837496 4840 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837500 4840 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837503 4840 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837507 4840 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837511 4840 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837514 4840 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837518 4840 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837521 4840 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837525 4840 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837533 4840 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837536 4840 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837540 4840 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837543 4840 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837546 4840 feature_gate.go:330] unrecognized feature gate: InsightsConfig Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837570 4840 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837574 4840 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837577 4840 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837581 4840 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837584 4840 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837588 4840 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837591 4840 feature_gate.go:330] unrecognized feature gate: OVNObservability Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837594 4840 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837598 4840 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837603 4840 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837607 4840 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837611 4840 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837616 4840 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837620 4840 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837624 4840 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837627 4840 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837631 4840 feature_gate.go:330] unrecognized feature gate: SignatureStores Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837634 4840 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837638 4840 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837643 4840 feature_gate.go:330] unrecognized feature gate: GatewayAPI Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837648 4840 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837653 4840 feature_gate.go:330] unrecognized feature gate: Example Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837657 4840 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837661 4840 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837665 4840 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837668 4840 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837672 4840 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837677 4840 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837681 4840 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837684 4840 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837688 4840 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837691 4840 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837695 4840 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837698 4840 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837702 4840 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837705 4840 feature_gate.go:330] unrecognized feature gate: PinnedImages Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837708 4840 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837712 4840 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837715 4840 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837720 4840 feature_gate.go:330] unrecognized feature gate: NewOLM Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837723 4840 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837726 4840 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837730 4840 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837734 4840 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837737 4840 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837741 4840 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837744 4840 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837748 4840 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837751 4840 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837759 4840 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837763 4840 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.837768 4840 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.837782 4840 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.846071 4840 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.846117 4840 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846189 4840 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846199 4840 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846206 4840 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846211 4840 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846215 4840 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846219 4840 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846223 4840 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846227 4840 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846266 4840 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846270 4840 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846274 4840 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846279 4840 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846284 4840 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846290 4840 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846295 4840 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846300 4840 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846304 4840 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846308 4840 feature_gate.go:330] unrecognized feature gate: SignatureStores Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846312 4840 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846316 4840 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846320 4840 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846324 4840 feature_gate.go:330] unrecognized feature gate: NewOLM Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846328 4840 feature_gate.go:330] unrecognized feature gate: GatewayAPI Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846331 4840 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846335 4840 feature_gate.go:330] unrecognized feature gate: OVNObservability Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846339 4840 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846344 4840 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846348 4840 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846352 4840 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846356 4840 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846360 4840 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846365 4840 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846368 4840 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846372 4840 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846391 4840 feature_gate.go:330] unrecognized feature gate: Example Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846401 4840 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846407 4840 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846411 4840 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846415 4840 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846419 4840 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846423 4840 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846426 4840 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846430 4840 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846433 4840 feature_gate.go:330] unrecognized feature gate: PlatformOperators Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846437 4840 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846441 4840 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846445 4840 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846449 4840 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846453 4840 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846457 4840 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846461 4840 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846464 4840 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846468 4840 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846471 4840 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846475 4840 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846479 4840 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846482 4840 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846486 4840 feature_gate.go:330] unrecognized feature gate: InsightsConfig Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846489 4840 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846493 4840 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846496 4840 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846500 4840 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846503 4840 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846507 4840 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846510 4840 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846514 4840 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846517 4840 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846521 4840 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846524 4840 feature_gate.go:330] unrecognized feature gate: PinnedImages Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846528 4840 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846532 4840 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.846540 4840 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846675 4840 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846682 4840 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846687 4840 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846691 4840 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846695 4840 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846699 4840 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846703 4840 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846707 4840 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846712 4840 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846715 4840 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846719 4840 feature_gate.go:330] unrecognized feature gate: PinnedImages Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846723 4840 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846726 4840 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846730 4840 feature_gate.go:330] unrecognized feature gate: SignatureStores Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846734 4840 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846738 4840 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846742 4840 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846746 4840 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846750 4840 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846753 4840 feature_gate.go:330] unrecognized feature gate: PlatformOperators Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846757 4840 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846781 4840 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846784 4840 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846788 4840 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846792 4840 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846797 4840 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846801 4840 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846805 4840 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846809 4840 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846813 4840 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846818 4840 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846822 4840 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846827 4840 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846832 4840 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846837 4840 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846841 4840 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846845 4840 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846849 4840 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846852 4840 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846856 4840 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846859 4840 feature_gate.go:330] unrecognized feature gate: GatewayAPI Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846863 4840 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846867 4840 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846871 4840 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846875 4840 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846879 4840 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846883 4840 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846889 4840 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846894 4840 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846898 4840 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846903 4840 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846907 4840 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846911 4840 feature_gate.go:330] unrecognized feature gate: OVNObservability Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846915 4840 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846919 4840 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846922 4840 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846926 4840 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846930 4840 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846934 4840 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846938 4840 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846941 4840 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846945 4840 feature_gate.go:330] unrecognized feature gate: Example Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846948 4840 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846952 4840 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846956 4840 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846959 4840 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846963 4840 feature_gate.go:330] unrecognized feature gate: InsightsConfig Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846967 4840 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846970 4840 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846974 4840 feature_gate.go:330] unrecognized feature gate: NewOLM Sep 30 13:56:11 crc kubenswrapper[4840]: W0930 13:56:11.846978 4840 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.846985 4840 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.847879 4840 server.go:940] "Client rotation is on, will bootstrap in background" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.852323 4840 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.852442 4840 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.853932 4840 server.go:997] "Starting client certificate rotation" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.853961 4840 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.854124 4840 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-18 03:30:25.823613278 +0000 UTC Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.854213 4840 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1885h34m13.969402698s for next certificate rotation Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.885519 4840 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.887997 4840 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.901814 4840 log.go:25] "Validated CRI v1 runtime API" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.939578 4840 log.go:25] "Validated CRI v1 image API" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.942106 4840 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.946662 4840 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-09-30-13-51-33-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.946695 4840 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:29 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:41 fsType:tmpfs blockSize:0}] Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.968782 4840 manager.go:217] Machine: {Timestamp:2025-09-30 13:56:11.964665567 +0000 UTC m=+0.593752030 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654120448 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:f41e0351-4eda-411a-9a99-43a1952f3d34 BootID:70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5 Filesystems:[{Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:41 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:29 Capacity:16827060224 Type:vfs Inodes:1048576 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:08:52:fe Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:08:52:fe Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:9e:56:1d Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:53:d8:b0 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:bb:60:66 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:f5:19:af Speed:-1 Mtu:1496} {Name:eth10 MacAddress:ba:29:e4:3a:52:e8 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:5e:a5:f3:7f:69:c1 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654120448 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.969055 4840 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.969295 4840 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.970616 4840 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.970846 4840 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.970889 4840 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.971113 4840 topology_manager.go:138] "Creating topology manager with none policy" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.971126 4840 container_manager_linux.go:303] "Creating device plugin manager" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.971860 4840 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.971891 4840 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.972132 4840 state_mem.go:36] "Initialized new in-memory state store" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.972223 4840 server.go:1245] "Using root directory" path="/var/lib/kubelet" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.988427 4840 kubelet.go:418] "Attempting to sync node with API server" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.988470 4840 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.988515 4840 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.988536 4840 kubelet.go:324] "Adding apiserver pod source" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.988583 4840 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 30 13:56:11 crc kubenswrapper[4840]: I0930 13:56:11.996403 4840 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.005816 4840 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Sep 30 13:56:12 crc kubenswrapper[4840]: W0930 13:56:12.011298 4840 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.122:6443: connect: connection refused Sep 30 13:56:12 crc kubenswrapper[4840]: E0930 13:56:12.011428 4840 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.122:6443: connect: connection refused" logger="UnhandledError" Sep 30 13:56:12 crc kubenswrapper[4840]: W0930 13:56:12.011299 4840 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.122:6443: connect: connection refused Sep 30 13:56:12 crc kubenswrapper[4840]: E0930 13:56:12.011491 4840 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.122:6443: connect: connection refused" logger="UnhandledError" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.014078 4840 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.023097 4840 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.023142 4840 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.023152 4840 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.023162 4840 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.023179 4840 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.023188 4840 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.023201 4840 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.023216 4840 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.023228 4840 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.023239 4840 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.023254 4840 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.023265 4840 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.023295 4840 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.023912 4840 server.go:1280] "Started kubelet" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.024108 4840 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.024223 4840 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.025534 4840 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 30 13:56:12 crc systemd[1]: Started Kubernetes Kubelet. Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.026087 4840 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.122:6443: connect: connection refused Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.029754 4840 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.029799 4840 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.029849 4840 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 12:48:52.531937049 +0000 UTC Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.029987 4840 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 2278h52m40.501953502s for next certificate rotation Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.029991 4840 volume_manager.go:287] "The desired_state_of_world populator starts" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.030007 4840 volume_manager.go:289] "Starting Kubelet Volume Manager" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.030095 4840 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Sep 30 13:56:12 crc kubenswrapper[4840]: E0930 13:56:12.030273 4840 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 30 13:56:12 crc kubenswrapper[4840]: W0930 13:56:12.040224 4840 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.122:6443: connect: connection refused Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.040844 4840 factory.go:55] Registering systemd factory Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.040903 4840 factory.go:221] Registration of the systemd container factory successfully Sep 30 13:56:12 crc kubenswrapper[4840]: E0930 13:56:12.040851 4840 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.122:6443: connect: connection refused" logger="UnhandledError" Sep 30 13:56:12 crc kubenswrapper[4840]: E0930 13:56:12.040774 4840 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.122:6443: connect: connection refused" interval="200ms" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.048640 4840 factory.go:153] Registering CRI-O factory Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.048696 4840 factory.go:221] Registration of the crio container factory successfully Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.048821 4840 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.048861 4840 factory.go:103] Registering Raw factory Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.048881 4840 manager.go:1196] Started watching for new ooms in manager Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.060769 4840 manager.go:319] Starting recovery of all containers Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.061720 4840 server.go:460] "Adding debug handlers to kubelet server" Sep 30 13:56:12 crc kubenswrapper[4840]: E0930 13:56:12.061492 4840 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.129.56.122:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186a13fab9f02ad1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-09-30 13:56:12.023876305 +0000 UTC m=+0.652962738,LastTimestamp:2025-09-30 13:56:12.023876305 +0000 UTC m=+0.652962738,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.066218 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.066662 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.066676 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.066688 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.066699 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.066709 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.066721 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.066733 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.066747 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.066758 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.066770 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.066784 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.066796 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.066813 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.066824 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.066837 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.066851 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.066864 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.066877 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.066890 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.066904 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.066921 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.066935 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.066947 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.066959 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.066973 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.066987 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067031 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067045 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067057 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067069 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067083 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067096 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067110 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067121 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067133 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067143 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067157 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067168 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067182 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067192 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067204 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067222 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067233 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067242 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067253 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067264 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067276 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067289 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067301 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067315 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067328 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067382 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067396 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067412 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067425 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067439 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067450 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067462 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067472 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067484 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067497 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067508 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067522 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067535 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067561 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067573 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067584 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067595 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067606 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067618 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067628 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067640 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067678 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067693 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067704 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067716 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067728 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067739 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067752 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067764 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067775 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067787 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067799 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067810 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067821 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067832 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067845 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067856 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067867 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067880 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067891 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067906 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067917 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067929 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067942 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067954 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067966 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067978 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.067990 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.068003 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.068015 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.068027 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.068039 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.068056 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.068072 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.068085 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.068097 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.068108 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.068120 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.068132 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.068146 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.068157 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.068169 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.068181 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.068192 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.068203 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.068214 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.069898 4840 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.069925 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.069939 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.069951 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.069968 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.069982 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.069996 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070009 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070021 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070080 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070095 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070108 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070122 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070166 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070180 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070194 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070246 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070274 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070288 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070301 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070342 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070357 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070394 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070407 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070418 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070473 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070487 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070501 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070514 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070532 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070544 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070571 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070583 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070677 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070689 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070701 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070713 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070731 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070764 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070776 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070790 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070858 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070873 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070886 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070938 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070950 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.070962 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071015 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071028 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071064 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071136 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071150 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071162 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071174 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071210 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071221 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071285 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071341 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071354 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071366 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071378 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071390 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071404 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071415 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071426 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071455 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071466 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071478 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071491 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071529 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071612 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071625 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071681 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071708 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071720 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071732 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071744 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071756 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071797 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071811 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071824 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071854 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071879 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071890 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071902 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071941 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071966 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071978 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.071989 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.072036 4840 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.072048 4840 reconstruct.go:97] "Volume reconstruction finished" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.072056 4840 reconciler.go:26] "Reconciler: start to sync state" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.084670 4840 manager.go:324] Recovery completed Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.098952 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.103810 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.103879 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.103898 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.104845 4840 cpu_manager.go:225] "Starting CPU manager" policy="none" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.104957 4840 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.105022 4840 state_mem.go:36] "Initialized new in-memory state store" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.112592 4840 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.115099 4840 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.115157 4840 status_manager.go:217] "Starting to sync pod status with apiserver" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.115188 4840 kubelet.go:2335] "Starting kubelet main sync loop" Sep 30 13:56:12 crc kubenswrapper[4840]: E0930 13:56:12.115245 4840 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 30 13:56:12 crc kubenswrapper[4840]: W0930 13:56:12.117031 4840 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.122:6443: connect: connection refused Sep 30 13:56:12 crc kubenswrapper[4840]: E0930 13:56:12.117139 4840 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.122:6443: connect: connection refused" logger="UnhandledError" Sep 30 13:56:12 crc kubenswrapper[4840]: E0930 13:56:12.131278 4840 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.155333 4840 policy_none.go:49] "None policy: Start" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.156857 4840 memory_manager.go:170] "Starting memorymanager" policy="None" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.156906 4840 state_mem.go:35] "Initializing new in-memory state store" Sep 30 13:56:12 crc kubenswrapper[4840]: E0930 13:56:12.216329 4840 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.225912 4840 manager.go:334] "Starting Device Plugin manager" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.225986 4840 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.226001 4840 server.go:79] "Starting device plugin registration server" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.226590 4840 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.226623 4840 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.226801 4840 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.226872 4840 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.226878 4840 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 30 13:56:12 crc kubenswrapper[4840]: E0930 13:56:12.236492 4840 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Sep 30 13:56:12 crc kubenswrapper[4840]: E0930 13:56:12.242603 4840 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.122:6443: connect: connection refused" interval="400ms" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.327296 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.329072 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.329142 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.329163 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.329209 4840 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 30 13:56:12 crc kubenswrapper[4840]: E0930 13:56:12.330157 4840 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.122:6443: connect: connection refused" node="crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.416607 4840 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc"] Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.416777 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.418326 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.418361 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.418374 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.418499 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.419412 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.419453 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.419471 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.419756 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.419813 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.420117 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.420178 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.420211 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.420955 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.420974 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.420983 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.421412 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.421457 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.421477 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.421461 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.421571 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.421583 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.421709 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.422007 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.422075 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.422685 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.422701 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.422712 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.422802 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.422981 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.423026 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.423109 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.423136 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.423151 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.423455 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.423471 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.423482 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.423632 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.423654 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.423802 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.423822 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.423832 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.424228 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.424248 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.424257 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.477983 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.478043 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.478075 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.478257 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.478310 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.478346 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.478371 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.478398 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.478426 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.478477 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.478504 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.478529 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.478574 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.478673 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.478740 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.530849 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.532993 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.533045 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.533058 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.533095 4840 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 30 13:56:12 crc kubenswrapper[4840]: E0930 13:56:12.533677 4840 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.122:6443: connect: connection refused" node="crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.580521 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.580903 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.581009 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.581067 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.581142 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.580759 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.581100 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.581306 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.581326 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.581343 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.581358 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.581375 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.581402 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.581436 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.581461 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.581462 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.581481 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.581502 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.581485 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.581533 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.581618 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.581655 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.581540 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.581588 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.581606 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.581542 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.581637 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.581803 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.581573 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.581874 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: E0930 13:56:12.643629 4840 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.122:6443: connect: connection refused" interval="800ms" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.745194 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.756161 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.782488 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.803038 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.809982 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 13:56:12 crc kubenswrapper[4840]: W0930 13:56:12.853115 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-bffe4ef6a83feb8626824e439b5b45a392734ef571b8dd78cd5404b4dbdf66a7 WatchSource:0}: Error finding container bffe4ef6a83feb8626824e439b5b45a392734ef571b8dd78cd5404b4dbdf66a7: Status 404 returned error can't find the container with id bffe4ef6a83feb8626824e439b5b45a392734ef571b8dd78cd5404b4dbdf66a7 Sep 30 13:56:12 crc kubenswrapper[4840]: W0930 13:56:12.861806 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-76b0a19152f70435055799e55bf4b713618ef6202ea4243659a0e007b2194d7a WatchSource:0}: Error finding container 76b0a19152f70435055799e55bf4b713618ef6202ea4243659a0e007b2194d7a: Status 404 returned error can't find the container with id 76b0a19152f70435055799e55bf4b713618ef6202ea4243659a0e007b2194d7a Sep 30 13:56:12 crc kubenswrapper[4840]: W0930 13:56:12.868297 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-3c223670f38c0adf1c7f04074ca51dfaec8f726bf8291d3cab3faa6b18ac92d6 WatchSource:0}: Error finding container 3c223670f38c0adf1c7f04074ca51dfaec8f726bf8291d3cab3faa6b18ac92d6: Status 404 returned error can't find the container with id 3c223670f38c0adf1c7f04074ca51dfaec8f726bf8291d3cab3faa6b18ac92d6 Sep 30 13:56:12 crc kubenswrapper[4840]: W0930 13:56:12.877453 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-01653bbdd1e3af9c07ebc012d59668c5f388f2ca3f4fd18adf434330209f950c WatchSource:0}: Error finding container 01653bbdd1e3af9c07ebc012d59668c5f388f2ca3f4fd18adf434330209f950c: Status 404 returned error can't find the container with id 01653bbdd1e3af9c07ebc012d59668c5f388f2ca3f4fd18adf434330209f950c Sep 30 13:56:12 crc kubenswrapper[4840]: W0930 13:56:12.878165 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-38a3a26bc045dbb25d692b945c55f718fd9b785adc01787bc1d68d1e29d72a04 WatchSource:0}: Error finding container 38a3a26bc045dbb25d692b945c55f718fd9b785adc01787bc1d68d1e29d72a04: Status 404 returned error can't find the container with id 38a3a26bc045dbb25d692b945c55f718fd9b785adc01787bc1d68d1e29d72a04 Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.934568 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.935866 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.935910 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.935924 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:12 crc kubenswrapper[4840]: I0930 13:56:12.935955 4840 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 30 13:56:12 crc kubenswrapper[4840]: E0930 13:56:12.936384 4840 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.122:6443: connect: connection refused" node="crc" Sep 30 13:56:13 crc kubenswrapper[4840]: I0930 13:56:13.027754 4840 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.122:6443: connect: connection refused Sep 30 13:56:13 crc kubenswrapper[4840]: W0930 13:56:13.043709 4840 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.122:6443: connect: connection refused Sep 30 13:56:13 crc kubenswrapper[4840]: E0930 13:56:13.043828 4840 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.122:6443: connect: connection refused" logger="UnhandledError" Sep 30 13:56:13 crc kubenswrapper[4840]: I0930 13:56:13.119364 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"76b0a19152f70435055799e55bf4b713618ef6202ea4243659a0e007b2194d7a"} Sep 30 13:56:13 crc kubenswrapper[4840]: I0930 13:56:13.120381 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"bffe4ef6a83feb8626824e439b5b45a392734ef571b8dd78cd5404b4dbdf66a7"} Sep 30 13:56:13 crc kubenswrapper[4840]: I0930 13:56:13.121668 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"38a3a26bc045dbb25d692b945c55f718fd9b785adc01787bc1d68d1e29d72a04"} Sep 30 13:56:13 crc kubenswrapper[4840]: I0930 13:56:13.122909 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"01653bbdd1e3af9c07ebc012d59668c5f388f2ca3f4fd18adf434330209f950c"} Sep 30 13:56:13 crc kubenswrapper[4840]: I0930 13:56:13.123958 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3c223670f38c0adf1c7f04074ca51dfaec8f726bf8291d3cab3faa6b18ac92d6"} Sep 30 13:56:13 crc kubenswrapper[4840]: W0930 13:56:13.280336 4840 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.122:6443: connect: connection refused Sep 30 13:56:13 crc kubenswrapper[4840]: E0930 13:56:13.280448 4840 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.122:6443: connect: connection refused" logger="UnhandledError" Sep 30 13:56:13 crc kubenswrapper[4840]: W0930 13:56:13.382513 4840 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.122:6443: connect: connection refused Sep 30 13:56:13 crc kubenswrapper[4840]: E0930 13:56:13.382618 4840 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.122:6443: connect: connection refused" logger="UnhandledError" Sep 30 13:56:13 crc kubenswrapper[4840]: E0930 13:56:13.444778 4840 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.122:6443: connect: connection refused" interval="1.6s" Sep 30 13:56:13 crc kubenswrapper[4840]: W0930 13:56:13.518335 4840 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.122:6443: connect: connection refused Sep 30 13:56:13 crc kubenswrapper[4840]: E0930 13:56:13.518435 4840 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.122:6443: connect: connection refused" logger="UnhandledError" Sep 30 13:56:13 crc kubenswrapper[4840]: I0930 13:56:13.736873 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:13 crc kubenswrapper[4840]: I0930 13:56:13.738358 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:13 crc kubenswrapper[4840]: I0930 13:56:13.738401 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:13 crc kubenswrapper[4840]: I0930 13:56:13.738418 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:13 crc kubenswrapper[4840]: I0930 13:56:13.738449 4840 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 30 13:56:13 crc kubenswrapper[4840]: E0930 13:56:13.738927 4840 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.122:6443: connect: connection refused" node="crc" Sep 30 13:56:14 crc kubenswrapper[4840]: I0930 13:56:14.027950 4840 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.122:6443: connect: connection refused Sep 30 13:56:15 crc kubenswrapper[4840]: I0930 13:56:15.027131 4840 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.122:6443: connect: connection refused Sep 30 13:56:15 crc kubenswrapper[4840]: E0930 13:56:15.045864 4840 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.122:6443: connect: connection refused" interval="3.2s" Sep 30 13:56:15 crc kubenswrapper[4840]: I0930 13:56:15.131900 4840 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="567d4e94a770ba3318053366878e777f03b826d75569ada21be35f83e4f21740" exitCode=0 Sep 30 13:56:15 crc kubenswrapper[4840]: I0930 13:56:15.132008 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:15 crc kubenswrapper[4840]: I0930 13:56:15.132005 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"567d4e94a770ba3318053366878e777f03b826d75569ada21be35f83e4f21740"} Sep 30 13:56:15 crc kubenswrapper[4840]: I0930 13:56:15.133064 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:15 crc kubenswrapper[4840]: I0930 13:56:15.133096 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:15 crc kubenswrapper[4840]: I0930 13:56:15.133107 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:15 crc kubenswrapper[4840]: I0930 13:56:15.133835 4840 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="08be626d75cb6bf53342c4ccab7abef9374735cd29d4afb7c7b5b62715db5d22" exitCode=0 Sep 30 13:56:15 crc kubenswrapper[4840]: I0930 13:56:15.133890 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:15 crc kubenswrapper[4840]: I0930 13:56:15.133919 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"08be626d75cb6bf53342c4ccab7abef9374735cd29d4afb7c7b5b62715db5d22"} Sep 30 13:56:15 crc kubenswrapper[4840]: I0930 13:56:15.134623 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:15 crc kubenswrapper[4840]: I0930 13:56:15.134657 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:15 crc kubenswrapper[4840]: I0930 13:56:15.134668 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:15 crc kubenswrapper[4840]: I0930 13:56:15.136759 4840 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8" exitCode=0 Sep 30 13:56:15 crc kubenswrapper[4840]: I0930 13:56:15.136818 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8"} Sep 30 13:56:15 crc kubenswrapper[4840]: I0930 13:56:15.136952 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:15 crc kubenswrapper[4840]: I0930 13:56:15.138165 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:15 crc kubenswrapper[4840]: I0930 13:56:15.138202 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:15 crc kubenswrapper[4840]: I0930 13:56:15.138215 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:15 crc kubenswrapper[4840]: I0930 13:56:15.140388 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3a5ae0096bcb76653145ceec82b8fdbac6c96f73d5a6f3fe96d6eb68698da809"} Sep 30 13:56:15 crc kubenswrapper[4840]: I0930 13:56:15.140421 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"28fce80735bf810d76d3e150ef69b57e14148c6ae09c2cf4da0cde1c95f01d13"} Sep 30 13:56:15 crc kubenswrapper[4840]: I0930 13:56:15.140438 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"547af187ef982f150029ae4ed1b0c730240906ef98a3879fbb63647383b15eb6"} Sep 30 13:56:15 crc kubenswrapper[4840]: I0930 13:56:15.141350 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:15 crc kubenswrapper[4840]: I0930 13:56:15.142108 4840 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="c5a84f046af7784b7ef2afb1b87d108573cd76d20f1b74077166f4dd7594f0b9" exitCode=0 Sep 30 13:56:15 crc kubenswrapper[4840]: I0930 13:56:15.142147 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"c5a84f046af7784b7ef2afb1b87d108573cd76d20f1b74077166f4dd7594f0b9"} Sep 30 13:56:15 crc kubenswrapper[4840]: I0930 13:56:15.142177 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:15 crc kubenswrapper[4840]: I0930 13:56:15.142198 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:15 crc kubenswrapper[4840]: I0930 13:56:15.142210 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:15 crc kubenswrapper[4840]: I0930 13:56:15.142200 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:15 crc kubenswrapper[4840]: I0930 13:56:15.142887 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:15 crc kubenswrapper[4840]: I0930 13:56:15.142912 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:15 crc kubenswrapper[4840]: I0930 13:56:15.142921 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:15 crc kubenswrapper[4840]: I0930 13:56:15.339678 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:15 crc kubenswrapper[4840]: I0930 13:56:15.340969 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:15 crc kubenswrapper[4840]: I0930 13:56:15.341023 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:15 crc kubenswrapper[4840]: I0930 13:56:15.341034 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:15 crc kubenswrapper[4840]: I0930 13:56:15.341065 4840 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 30 13:56:15 crc kubenswrapper[4840]: E0930 13:56:15.341853 4840 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.122:6443: connect: connection refused" node="crc" Sep 30 13:56:15 crc kubenswrapper[4840]: W0930 13:56:15.487360 4840 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.122:6443: connect: connection refused Sep 30 13:56:15 crc kubenswrapper[4840]: E0930 13:56:15.487439 4840 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.122:6443: connect: connection refused" logger="UnhandledError" Sep 30 13:56:15 crc kubenswrapper[4840]: W0930 13:56:15.741158 4840 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.122:6443: connect: connection refused Sep 30 13:56:15 crc kubenswrapper[4840]: E0930 13:56:15.741256 4840 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.122:6443: connect: connection refused" logger="UnhandledError" Sep 30 13:56:15 crc kubenswrapper[4840]: W0930 13:56:15.882544 4840 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.122:6443: connect: connection refused Sep 30 13:56:15 crc kubenswrapper[4840]: E0930 13:56:15.882668 4840 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.122:6443: connect: connection refused" logger="UnhandledError" Sep 30 13:56:16 crc kubenswrapper[4840]: I0930 13:56:16.027579 4840 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.122:6443: connect: connection refused Sep 30 13:56:16 crc kubenswrapper[4840]: I0930 13:56:16.147070 4840 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="7c2c3a885cd937e387b3ba820ef501fe0c318a277d2353aa6664b3674faf79de" exitCode=0 Sep 30 13:56:16 crc kubenswrapper[4840]: I0930 13:56:16.147528 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"7c2c3a885cd937e387b3ba820ef501fe0c318a277d2353aa6664b3674faf79de"} Sep 30 13:56:16 crc kubenswrapper[4840]: I0930 13:56:16.151134 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:16 crc kubenswrapper[4840]: I0930 13:56:16.154735 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:16 crc kubenswrapper[4840]: I0930 13:56:16.154781 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:16 crc kubenswrapper[4840]: I0930 13:56:16.154800 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:16 crc kubenswrapper[4840]: I0930 13:56:16.155121 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"2c8399ebe59b7663b0af7b042b0b4e67fcdf61efe10c803f50b1184fce406456"} Sep 30 13:56:16 crc kubenswrapper[4840]: I0930 13:56:16.155273 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:16 crc kubenswrapper[4840]: I0930 13:56:16.156456 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:16 crc kubenswrapper[4840]: I0930 13:56:16.156582 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:16 crc kubenswrapper[4840]: I0930 13:56:16.156599 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:16 crc kubenswrapper[4840]: I0930 13:56:16.159205 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"af3f57df6d22c47163751c297aff0ca1dc2bdbf150b9e385e72acdefde92910c"} Sep 30 13:56:16 crc kubenswrapper[4840]: I0930 13:56:16.159244 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"9a08b53dbed5f3a033d2838519733a2a4109ac5a954d9f5fabd0762b2823b090"} Sep 30 13:56:16 crc kubenswrapper[4840]: I0930 13:56:16.159252 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:16 crc kubenswrapper[4840]: I0930 13:56:16.159281 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"8cf852cfa87ef328d172369c26197a9e6a58fe26fa3113469d9275b2e78c30c2"} Sep 30 13:56:16 crc kubenswrapper[4840]: I0930 13:56:16.160132 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:16 crc kubenswrapper[4840]: I0930 13:56:16.160191 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:16 crc kubenswrapper[4840]: I0930 13:56:16.160205 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:16 crc kubenswrapper[4840]: I0930 13:56:16.162944 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8c507106f39ae2e8acd4268c4e8f28ffd88a0965e69544d8d8a8830e83eceaf9"} Sep 30 13:56:16 crc kubenswrapper[4840]: I0930 13:56:16.162995 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3d62fbefb67a24fedc5f13652596c76f5b288fec58d79b5209f49c79db139d79"} Sep 30 13:56:16 crc kubenswrapper[4840]: I0930 13:56:16.163007 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"cba320f443ed90a4e755228b663a1da3096e7ea8e068671d1f6273544b6ef60b"} Sep 30 13:56:16 crc kubenswrapper[4840]: I0930 13:56:16.165798 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"637b464972cfd0cd6f1f34448bf1dc14cf929126083bfc9081664fa6b25cd743"} Sep 30 13:56:16 crc kubenswrapper[4840]: I0930 13:56:16.165967 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:16 crc kubenswrapper[4840]: I0930 13:56:16.166933 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:16 crc kubenswrapper[4840]: I0930 13:56:16.166981 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:16 crc kubenswrapper[4840]: I0930 13:56:16.167001 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:16 crc kubenswrapper[4840]: W0930 13:56:16.240140 4840 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.122:6443: connect: connection refused Sep 30 13:56:16 crc kubenswrapper[4840]: E0930 13:56:16.240250 4840 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.122:6443: connect: connection refused" logger="UnhandledError" Sep 30 13:56:17 crc kubenswrapper[4840]: I0930 13:56:17.025591 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 13:56:17 crc kubenswrapper[4840]: I0930 13:56:17.027925 4840 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.122:6443: connect: connection refused Sep 30 13:56:17 crc kubenswrapper[4840]: I0930 13:56:17.170886 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"350358b446502637fedd26f8a703dcbba37af9f66edaeec55dd4e7323cc24945"} Sep 30 13:56:17 crc kubenswrapper[4840]: I0930 13:56:17.170932 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f1435e2e88d45e5561ff5863095d0f953ef1478d19d87691ab226cad7ae148f7"} Sep 30 13:56:17 crc kubenswrapper[4840]: I0930 13:56:17.171023 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:17 crc kubenswrapper[4840]: I0930 13:56:17.172179 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:17 crc kubenswrapper[4840]: I0930 13:56:17.172210 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:17 crc kubenswrapper[4840]: I0930 13:56:17.172221 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:17 crc kubenswrapper[4840]: I0930 13:56:17.173015 4840 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="f335b20cc833d3278d09697f85b1abd85732aefd80429fb1866eca5565ef98e4" exitCode=0 Sep 30 13:56:17 crc kubenswrapper[4840]: I0930 13:56:17.173110 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:17 crc kubenswrapper[4840]: I0930 13:56:17.173586 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:17 crc kubenswrapper[4840]: I0930 13:56:17.173842 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"f335b20cc833d3278d09697f85b1abd85732aefd80429fb1866eca5565ef98e4"} Sep 30 13:56:17 crc kubenswrapper[4840]: I0930 13:56:17.173903 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:17 crc kubenswrapper[4840]: I0930 13:56:17.174234 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:17 crc kubenswrapper[4840]: I0930 13:56:17.174514 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:17 crc kubenswrapper[4840]: I0930 13:56:17.174539 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:17 crc kubenswrapper[4840]: I0930 13:56:17.174565 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:17 crc kubenswrapper[4840]: I0930 13:56:17.174581 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:17 crc kubenswrapper[4840]: I0930 13:56:17.174593 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:17 crc kubenswrapper[4840]: I0930 13:56:17.174602 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:17 crc kubenswrapper[4840]: I0930 13:56:17.174814 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:17 crc kubenswrapper[4840]: I0930 13:56:17.174864 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:17 crc kubenswrapper[4840]: I0930 13:56:17.174883 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:17 crc kubenswrapper[4840]: I0930 13:56:17.174910 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:17 crc kubenswrapper[4840]: I0930 13:56:17.174889 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:17 crc kubenswrapper[4840]: I0930 13:56:17.174976 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:17 crc kubenswrapper[4840]: I0930 13:56:17.550373 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 13:56:18 crc kubenswrapper[4840]: I0930 13:56:18.027956 4840 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.122:6443: connect: connection refused Sep 30 13:56:18 crc kubenswrapper[4840]: I0930 13:56:18.188746 4840 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 13:56:18 crc kubenswrapper[4840]: I0930 13:56:18.188800 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:18 crc kubenswrapper[4840]: I0930 13:56:18.188816 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:18 crc kubenswrapper[4840]: I0930 13:56:18.189399 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"987b6a297870d2898fe2b5df90b15166fb192e10d232fa9bae686ebb7f4180aa"} Sep 30 13:56:18 crc kubenswrapper[4840]: I0930 13:56:18.189439 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"15be3c94974cfafcf69a74ef1ff8f6b9da74840805ceafb96aeb903dbf90fd58"} Sep 30 13:56:18 crc kubenswrapper[4840]: I0930 13:56:18.189450 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d68b17f8a9e41caf54d2f8e2ae775c33d653299e8fe157de0fa4596e1339682e"} Sep 30 13:56:18 crc kubenswrapper[4840]: I0930 13:56:18.189592 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:18 crc kubenswrapper[4840]: I0930 13:56:18.189906 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:18 crc kubenswrapper[4840]: I0930 13:56:18.189946 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:18 crc kubenswrapper[4840]: I0930 13:56:18.189959 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:18 crc kubenswrapper[4840]: I0930 13:56:18.190609 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:18 crc kubenswrapper[4840]: I0930 13:56:18.190651 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:18 crc kubenswrapper[4840]: I0930 13:56:18.190665 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:18 crc kubenswrapper[4840]: I0930 13:56:18.192031 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:18 crc kubenswrapper[4840]: I0930 13:56:18.192050 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:18 crc kubenswrapper[4840]: I0930 13:56:18.192059 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:18 crc kubenswrapper[4840]: E0930 13:56:18.247188 4840 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.122:6443: connect: connection refused" interval="6.4s" Sep 30 13:56:18 crc kubenswrapper[4840]: I0930 13:56:18.542728 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:18 crc kubenswrapper[4840]: I0930 13:56:18.544136 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:18 crc kubenswrapper[4840]: I0930 13:56:18.544169 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:18 crc kubenswrapper[4840]: I0930 13:56:18.544179 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:18 crc kubenswrapper[4840]: I0930 13:56:18.544206 4840 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 30 13:56:18 crc kubenswrapper[4840]: I0930 13:56:18.969563 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 13:56:19 crc kubenswrapper[4840]: I0930 13:56:19.194079 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Sep 30 13:56:19 crc kubenswrapper[4840]: I0930 13:56:19.197857 4840 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="350358b446502637fedd26f8a703dcbba37af9f66edaeec55dd4e7323cc24945" exitCode=255 Sep 30 13:56:19 crc kubenswrapper[4840]: I0930 13:56:19.197898 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"350358b446502637fedd26f8a703dcbba37af9f66edaeec55dd4e7323cc24945"} Sep 30 13:56:19 crc kubenswrapper[4840]: I0930 13:56:19.198088 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:19 crc kubenswrapper[4840]: I0930 13:56:19.200124 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:19 crc kubenswrapper[4840]: I0930 13:56:19.200178 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:19 crc kubenswrapper[4840]: I0930 13:56:19.200205 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:19 crc kubenswrapper[4840]: I0930 13:56:19.200983 4840 scope.go:117] "RemoveContainer" containerID="350358b446502637fedd26f8a703dcbba37af9f66edaeec55dd4e7323cc24945" Sep 30 13:56:19 crc kubenswrapper[4840]: I0930 13:56:19.205839 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"15291f77e6b50679df8feef2f536b5f5745d6f841372770d647831006d8eb0af"} Sep 30 13:56:19 crc kubenswrapper[4840]: I0930 13:56:19.205866 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"87a3cd4212aed96f1cad501a5c62065b9c1a289c03afbec269ca193c12434c56"} Sep 30 13:56:19 crc kubenswrapper[4840]: I0930 13:56:19.206041 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:19 crc kubenswrapper[4840]: I0930 13:56:19.207232 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:19 crc kubenswrapper[4840]: I0930 13:56:19.207269 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:19 crc kubenswrapper[4840]: I0930 13:56:19.207284 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:19 crc kubenswrapper[4840]: I0930 13:56:19.923738 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 13:56:19 crc kubenswrapper[4840]: I0930 13:56:19.923966 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:19 crc kubenswrapper[4840]: I0930 13:56:19.925183 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:19 crc kubenswrapper[4840]: I0930 13:56:19.925214 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:19 crc kubenswrapper[4840]: I0930 13:56:19.925226 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:19 crc kubenswrapper[4840]: I0930 13:56:19.930987 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 13:56:20 crc kubenswrapper[4840]: I0930 13:56:20.209991 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Sep 30 13:56:20 crc kubenswrapper[4840]: I0930 13:56:20.211877 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb"} Sep 30 13:56:20 crc kubenswrapper[4840]: I0930 13:56:20.211961 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:20 crc kubenswrapper[4840]: I0930 13:56:20.212059 4840 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 13:56:20 crc kubenswrapper[4840]: I0930 13:56:20.212148 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:20 crc kubenswrapper[4840]: I0930 13:56:20.212469 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:20 crc kubenswrapper[4840]: I0930 13:56:20.213270 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:20 crc kubenswrapper[4840]: I0930 13:56:20.213276 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:20 crc kubenswrapper[4840]: I0930 13:56:20.213321 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:20 crc kubenswrapper[4840]: I0930 13:56:20.213330 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:20 crc kubenswrapper[4840]: I0930 13:56:20.213340 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:20 crc kubenswrapper[4840]: I0930 13:56:20.213340 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:20 crc kubenswrapper[4840]: I0930 13:56:20.213402 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:20 crc kubenswrapper[4840]: I0930 13:56:20.213506 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:20 crc kubenswrapper[4840]: I0930 13:56:20.213525 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:20 crc kubenswrapper[4840]: I0930 13:56:20.310472 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Sep 30 13:56:21 crc kubenswrapper[4840]: I0930 13:56:21.215234 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:21 crc kubenswrapper[4840]: I0930 13:56:21.215247 4840 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 13:56:21 crc kubenswrapper[4840]: I0930 13:56:21.215376 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:21 crc kubenswrapper[4840]: I0930 13:56:21.216364 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:21 crc kubenswrapper[4840]: I0930 13:56:21.216403 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:21 crc kubenswrapper[4840]: I0930 13:56:21.216424 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:21 crc kubenswrapper[4840]: I0930 13:56:21.216434 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:21 crc kubenswrapper[4840]: I0930 13:56:21.216454 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:21 crc kubenswrapper[4840]: I0930 13:56:21.216463 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:21 crc kubenswrapper[4840]: I0930 13:56:21.560440 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 13:56:21 crc kubenswrapper[4840]: I0930 13:56:21.659525 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 13:56:22 crc kubenswrapper[4840]: I0930 13:56:22.217949 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:22 crc kubenswrapper[4840]: I0930 13:56:22.218928 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:22 crc kubenswrapper[4840]: I0930 13:56:22.218976 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:22 crc kubenswrapper[4840]: I0930 13:56:22.218993 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:22 crc kubenswrapper[4840]: E0930 13:56:22.236630 4840 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Sep 30 13:56:22 crc kubenswrapper[4840]: I0930 13:56:22.453362 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 13:56:22 crc kubenswrapper[4840]: I0930 13:56:22.453965 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:22 crc kubenswrapper[4840]: I0930 13:56:22.455672 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:22 crc kubenswrapper[4840]: I0930 13:56:22.455728 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:22 crc kubenswrapper[4840]: I0930 13:56:22.455747 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:23 crc kubenswrapper[4840]: I0930 13:56:23.221217 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:23 crc kubenswrapper[4840]: I0930 13:56:23.222542 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:23 crc kubenswrapper[4840]: I0930 13:56:23.222616 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:23 crc kubenswrapper[4840]: I0930 13:56:23.222628 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:23 crc kubenswrapper[4840]: I0930 13:56:23.663788 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 13:56:23 crc kubenswrapper[4840]: I0930 13:56:23.663952 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:23 crc kubenswrapper[4840]: I0930 13:56:23.665089 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:23 crc kubenswrapper[4840]: I0930 13:56:23.665126 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:23 crc kubenswrapper[4840]: I0930 13:56:23.665151 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:25 crc kubenswrapper[4840]: I0930 13:56:25.453418 4840 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Sep 30 13:56:25 crc kubenswrapper[4840]: I0930 13:56:25.453505 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 30 13:56:27 crc kubenswrapper[4840]: I0930 13:56:27.556184 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 13:56:27 crc kubenswrapper[4840]: I0930 13:56:27.556316 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:27 crc kubenswrapper[4840]: I0930 13:56:27.557495 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:27 crc kubenswrapper[4840]: I0930 13:56:27.557547 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:27 crc kubenswrapper[4840]: I0930 13:56:27.557581 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:27 crc kubenswrapper[4840]: I0930 13:56:27.936029 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Sep 30 13:56:27 crc kubenswrapper[4840]: I0930 13:56:27.936247 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:27 crc kubenswrapper[4840]: I0930 13:56:27.940177 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:27 crc kubenswrapper[4840]: I0930 13:56:27.940238 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:27 crc kubenswrapper[4840]: I0930 13:56:27.940257 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:27 crc kubenswrapper[4840]: I0930 13:56:27.978823 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Sep 30 13:56:28 crc kubenswrapper[4840]: I0930 13:56:28.233689 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:28 crc kubenswrapper[4840]: I0930 13:56:28.234665 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:28 crc kubenswrapper[4840]: I0930 13:56:28.234703 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:28 crc kubenswrapper[4840]: I0930 13:56:28.234716 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:28 crc kubenswrapper[4840]: I0930 13:56:28.245319 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Sep 30 13:56:28 crc kubenswrapper[4840]: E0930 13:56:28.545908 4840 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Sep 30 13:56:28 crc kubenswrapper[4840]: I0930 13:56:28.969921 4840 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="Get \"https://192.168.126.11:6443/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Sep 30 13:56:28 crc kubenswrapper[4840]: I0930 13:56:28.970041 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Sep 30 13:56:29 crc kubenswrapper[4840]: I0930 13:56:29.028252 4840 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Sep 30 13:56:29 crc kubenswrapper[4840]: I0930 13:56:29.236804 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:29 crc kubenswrapper[4840]: I0930 13:56:29.238583 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:29 crc kubenswrapper[4840]: I0930 13:56:29.238625 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:29 crc kubenswrapper[4840]: I0930 13:56:29.238642 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:29 crc kubenswrapper[4840]: W0930 13:56:29.449170 4840 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Sep 30 13:56:29 crc kubenswrapper[4840]: I0930 13:56:29.449339 4840 trace.go:236] Trace[1023050918]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (30-Sep-2025 13:56:19.448) (total time: 10001ms): Sep 30 13:56:29 crc kubenswrapper[4840]: Trace[1023050918]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (13:56:29.449) Sep 30 13:56:29 crc kubenswrapper[4840]: Trace[1023050918]: [10.00124898s] [10.00124898s] END Sep 30 13:56:29 crc kubenswrapper[4840]: E0930 13:56:29.449391 4840 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Sep 30 13:56:29 crc kubenswrapper[4840]: I0930 13:56:29.636068 4840 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Sep 30 13:56:29 crc kubenswrapper[4840]: I0930 13:56:29.636125 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Sep 30 13:56:31 crc kubenswrapper[4840]: I0930 13:56:31.257249 4840 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Sep 30 13:56:31 crc kubenswrapper[4840]: I0930 13:56:31.257324 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Sep 30 13:56:31 crc kubenswrapper[4840]: I0930 13:56:31.660459 4840 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Sep 30 13:56:31 crc kubenswrapper[4840]: I0930 13:56:31.660534 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Sep 30 13:56:32 crc kubenswrapper[4840]: E0930 13:56:32.236773 4840 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Sep 30 13:56:33 crc kubenswrapper[4840]: I0930 13:56:33.973064 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 13:56:33 crc kubenswrapper[4840]: I0930 13:56:33.973286 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:33 crc kubenswrapper[4840]: I0930 13:56:33.973737 4840 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Sep 30 13:56:33 crc kubenswrapper[4840]: I0930 13:56:33.973802 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Sep 30 13:56:33 crc kubenswrapper[4840]: I0930 13:56:33.974391 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:33 crc kubenswrapper[4840]: I0930 13:56:33.974427 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:33 crc kubenswrapper[4840]: I0930 13:56:33.974440 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:33 crc kubenswrapper[4840]: I0930 13:56:33.976495 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.249471 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.249918 4840 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.249991 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.250369 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.250405 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.250418 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.629217 4840 trace.go:236] Trace[28936495]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (30-Sep-2025 13:56:19.917) (total time: 14711ms): Sep 30 13:56:34 crc kubenswrapper[4840]: Trace[28936495]: ---"Objects listed" error: 14711ms (13:56:34.629) Sep 30 13:56:34 crc kubenswrapper[4840]: Trace[28936495]: [14.71165151s] [14.71165151s] END Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.629246 4840 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.632147 4840 trace.go:236] Trace[2022167721]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (30-Sep-2025 13:56:19.810) (total time: 14822ms): Sep 30 13:56:34 crc kubenswrapper[4840]: Trace[2022167721]: ---"Objects listed" error: 14822ms (13:56:34.632) Sep 30 13:56:34 crc kubenswrapper[4840]: Trace[2022167721]: [14.822064243s] [14.822064243s] END Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.632179 4840 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.635723 4840 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.636594 4840 trace.go:236] Trace[950419586]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (30-Sep-2025 13:56:21.588) (total time: 13047ms): Sep 30 13:56:34 crc kubenswrapper[4840]: Trace[950419586]: ---"Objects listed" error: 13047ms (13:56:34.636) Sep 30 13:56:34 crc kubenswrapper[4840]: Trace[950419586]: [13.047813167s] [13.047813167s] END Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.636640 4840 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.804372 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.804587 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.805732 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.805764 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.805775 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.808594 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.946409 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.947635 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.947683 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.947696 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.947838 4840 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.956483 4840 kubelet_node_status.go:115] "Node was previously registered" node="crc" Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.956777 4840 kubelet_node_status.go:79] "Successfully registered node" node="crc" Sep 30 13:56:34 crc kubenswrapper[4840]: E0930 13:56:34.956811 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.963055 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.963101 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.963117 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.963141 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.963156 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:34Z","lastTransitionTime":"2025-09-30T13:56:34Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Sep 30 13:56:34 crc kubenswrapper[4840]: E0930 13:56:34.977430 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5\\\",\\\"systemUUID\\\":\\\"f41e0351-4eda-411a-9a99-43a1952f3d34\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.981920 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.981963 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.981975 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.981998 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.982010 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:34Z","lastTransitionTime":"2025-09-30T13:56:34Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Sep 30 13:56:34 crc kubenswrapper[4840]: E0930 13:56:34.990586 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5\\\",\\\"systemUUID\\\":\\\"f41e0351-4eda-411a-9a99-43a1952f3d34\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.994599 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.994651 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.994664 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.994691 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:34 crc kubenswrapper[4840]: I0930 13:56:34.994705 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:34Z","lastTransitionTime":"2025-09-30T13:56:34Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Sep 30 13:56:35 crc kubenswrapper[4840]: E0930 13:56:35.004839 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5\\\",\\\"systemUUID\\\":\\\"f41e0351-4eda-411a-9a99-43a1952f3d34\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 13:56:35 crc kubenswrapper[4840]: I0930 13:56:35.007924 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:35 crc kubenswrapper[4840]: I0930 13:56:35.007958 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:35 crc kubenswrapper[4840]: I0930 13:56:35.007968 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:35 crc kubenswrapper[4840]: I0930 13:56:35.007986 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:35 crc kubenswrapper[4840]: I0930 13:56:35.007996 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:35Z","lastTransitionTime":"2025-09-30T13:56:35Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Sep 30 13:56:35 crc kubenswrapper[4840]: E0930 13:56:35.017292 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:35Z\\\",\\\"message\\\":\\\"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5\\\",\\\"systemUUID\\\":\\\"f41e0351-4eda-411a-9a99-43a1952f3d34\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 13:56:35 crc kubenswrapper[4840]: E0930 13:56:35.017458 4840 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 30 13:56:35 crc kubenswrapper[4840]: E0930 13:56:35.017492 4840 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 30 13:56:35 crc kubenswrapper[4840]: E0930 13:56:35.117719 4840 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 30 13:56:35 crc kubenswrapper[4840]: E0930 13:56:35.218226 4840 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 30 13:56:35 crc kubenswrapper[4840]: I0930 13:56:35.253302 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Sep 30 13:56:35 crc kubenswrapper[4840]: I0930 13:56:35.253938 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Sep 30 13:56:35 crc kubenswrapper[4840]: I0930 13:56:35.255532 4840 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb" exitCode=255 Sep 30 13:56:35 crc kubenswrapper[4840]: I0930 13:56:35.255616 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb"} Sep 30 13:56:35 crc kubenswrapper[4840]: I0930 13:56:35.255672 4840 scope.go:117] "RemoveContainer" containerID="350358b446502637fedd26f8a703dcbba37af9f66edaeec55dd4e7323cc24945" Sep 30 13:56:35 crc kubenswrapper[4840]: I0930 13:56:35.255696 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:35 crc kubenswrapper[4840]: I0930 13:56:35.255746 4840 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 13:56:35 crc kubenswrapper[4840]: I0930 13:56:35.256945 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:35 crc kubenswrapper[4840]: I0930 13:56:35.256979 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:35 crc kubenswrapper[4840]: I0930 13:56:35.256991 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:35 crc kubenswrapper[4840]: I0930 13:56:35.258006 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:35 crc kubenswrapper[4840]: I0930 13:56:35.258031 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:35 crc kubenswrapper[4840]: I0930 13:56:35.258045 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:35 crc kubenswrapper[4840]: I0930 13:56:35.258470 4840 scope.go:117] "RemoveContainer" containerID="fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb" Sep 30 13:56:35 crc kubenswrapper[4840]: E0930 13:56:35.258671 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Sep 30 13:56:35 crc kubenswrapper[4840]: E0930 13:56:35.318341 4840 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 30 13:56:35 crc kubenswrapper[4840]: E0930 13:56:35.419202 4840 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 30 13:56:35 crc kubenswrapper[4840]: E0930 13:56:35.519886 4840 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 30 13:56:35 crc kubenswrapper[4840]: E0930 13:56:35.620996 4840 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 30 13:56:35 crc kubenswrapper[4840]: E0930 13:56:35.721761 4840 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 30 13:56:35 crc kubenswrapper[4840]: E0930 13:56:35.822389 4840 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 30 13:56:35 crc kubenswrapper[4840]: E0930 13:56:35.923147 4840 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 30 13:56:36 crc kubenswrapper[4840]: E0930 13:56:36.024296 4840 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 30 13:56:36 crc kubenswrapper[4840]: E0930 13:56:36.125340 4840 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 30 13:56:36 crc kubenswrapper[4840]: E0930 13:56:36.226019 4840 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 30 13:56:36 crc kubenswrapper[4840]: I0930 13:56:36.260151 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Sep 30 13:56:36 crc kubenswrapper[4840]: E0930 13:56:36.326880 4840 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 30 13:56:36 crc kubenswrapper[4840]: E0930 13:56:36.427804 4840 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 30 13:56:36 crc kubenswrapper[4840]: E0930 13:56:36.528855 4840 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 30 13:56:36 crc kubenswrapper[4840]: E0930 13:56:36.629502 4840 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 30 13:56:36 crc kubenswrapper[4840]: E0930 13:56:36.730094 4840 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 30 13:56:36 crc kubenswrapper[4840]: E0930 13:56:36.830455 4840 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 30 13:56:36 crc kubenswrapper[4840]: E0930 13:56:36.931361 4840 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 30 13:56:37 crc kubenswrapper[4840]: E0930 13:56:37.031468 4840 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 30 13:56:37 crc kubenswrapper[4840]: E0930 13:56:37.132334 4840 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 30 13:56:37 crc kubenswrapper[4840]: E0930 13:56:37.232505 4840 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 30 13:56:37 crc kubenswrapper[4840]: E0930 13:56:37.333213 4840 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 30 13:56:37 crc kubenswrapper[4840]: E0930 13:56:37.434215 4840 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 30 13:56:37 crc kubenswrapper[4840]: E0930 13:56:37.534973 4840 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 30 13:56:37 crc kubenswrapper[4840]: E0930 13:56:37.635411 4840 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 30 13:56:37 crc kubenswrapper[4840]: E0930 13:56:37.736069 4840 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 30 13:56:37 crc kubenswrapper[4840]: E0930 13:56:37.836613 4840 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 30 13:56:37 crc kubenswrapper[4840]: E0930 13:56:37.937563 4840 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.023887 4840 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.039715 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.039750 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.039759 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.039776 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.039787 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:38Z","lastTransitionTime":"2025-09-30T13:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.142409 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.142481 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.142505 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.142536 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.142601 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:38Z","lastTransitionTime":"2025-09-30T13:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.245730 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.245763 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.245773 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.245790 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.245801 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:38Z","lastTransitionTime":"2025-09-30T13:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.348046 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.348095 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.348108 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.348128 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.348140 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:38Z","lastTransitionTime":"2025-09-30T13:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.451107 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.451147 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.451156 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.451171 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.451180 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:38Z","lastTransitionTime":"2025-09-30T13:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.553750 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.553812 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.553828 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.553852 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.553868 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:38Z","lastTransitionTime":"2025-09-30T13:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.657390 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.657425 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.657435 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.657450 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.657460 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:38Z","lastTransitionTime":"2025-09-30T13:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.760494 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.760585 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.760602 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.760628 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.760644 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:38Z","lastTransitionTime":"2025-09-30T13:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.863710 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.863761 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.863776 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.863798 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.863818 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:38Z","lastTransitionTime":"2025-09-30T13:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.967352 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.967400 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.967417 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.967444 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:38 crc kubenswrapper[4840]: I0930 13:56:38.967458 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:38Z","lastTransitionTime":"2025-09-30T13:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.005418 4840 apiserver.go:52] "Watching apiserver" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.010359 4840 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.010755 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-bwvl2","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-node-identity/network-node-identity-vrzqb","openshift-ovn-kubernetes/ovnkube-node-kfx69","openshift-dns/node-resolver-w988d","openshift-machine-config-operator/machine-config-daemon-747gk","openshift-multus/multus-additional-cni-plugins-hmwzh","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-target-xd92c"] Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.011171 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.012025 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-747gk" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.012118 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.012026 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.012388 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.012416 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.012445 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 30 13:56:39 crc kubenswrapper[4840]: E0930 13:56:39.012432 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.012408 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.012525 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.012820 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.012882 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-w988d" Sep 30 13:56:39 crc kubenswrapper[4840]: E0930 13:56:39.013122 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:56:39 crc kubenswrapper[4840]: E0930 13:56:39.012413 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.014724 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.015179 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.018256 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.018318 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.018255 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.018489 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.018650 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.018860 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.018964 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.019092 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.019399 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.019609 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.019673 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.019721 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.019622 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.019915 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.020032 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.020387 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.020508 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.020877 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.021009 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.021071 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.021032 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.021622 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.021668 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.021908 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.021909 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.022192 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.025283 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.025719 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.026022 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.040050 4840 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.046616 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.063931 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.065889 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.065926 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.065951 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.065971 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.065989 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066008 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066025 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066045 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066065 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066083 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066098 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066114 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066131 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066148 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066165 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066184 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066203 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066222 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066237 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066252 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066248 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066268 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066284 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066300 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066315 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066331 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066360 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066394 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066409 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066429 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066447 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066454 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066472 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066489 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066505 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066520 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066534 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066563 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066626 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066570 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066645 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066661 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066679 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066718 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066741 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066829 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066845 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066861 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066877 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066894 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066910 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066925 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066939 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066956 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.066986 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.067013 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.067018 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.067052 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.067068 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.067082 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.067099 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.067114 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.067135 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.067168 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.067212 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.067160 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.067270 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.067289 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.067304 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.067320 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.067336 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.067380 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.067467 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.067513 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.067564 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.067587 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.067614 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.067633 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.067651 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.067668 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.067683 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.067701 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.067729 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.067751 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.067774 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.067796 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.067824 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.067846 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.067867 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.067902 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.067978 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.068076 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.068102 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.068124 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.068150 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.068175 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.068207 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.068230 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.068255 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.068261 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.068279 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.068302 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.068325 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.068348 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.068373 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.068397 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.068421 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.068447 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.068472 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.068498 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.068522 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.068566 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.068591 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.068614 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.068637 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.068669 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.068695 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.068721 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.068748 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.068771 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.068794 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.068813 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.068819 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.068891 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.068901 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.068919 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.068952 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.068974 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069001 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069070 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069088 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069109 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069103 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069128 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069128 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069147 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069170 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069152 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069186 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069204 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069237 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069254 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069307 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069325 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069344 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069364 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069386 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069404 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069423 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069441 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069458 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069476 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069528 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069571 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069591 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069610 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069630 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069648 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069673 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069692 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069709 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069734 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069751 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069768 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069787 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069804 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069822 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069838 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069857 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069884 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069907 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069935 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069958 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069980 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069998 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070016 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070034 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070050 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070070 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070087 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070106 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070124 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070143 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070163 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070184 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070202 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070220 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070236 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070254 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070326 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070350 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070367 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070386 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070405 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070422 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070442 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070458 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070479 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070498 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070517 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070535 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070572 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070592 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070610 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070628 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070649 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070667 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070685 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070704 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070721 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070738 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070800 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/da1d702d-bbe7-4e02-aa18-cb6556383674-hosts-file\") pod \"node-resolver-w988d\" (UID: \"da1d702d-bbe7-4e02-aa18-cb6556383674\") " pod="openshift-dns/node-resolver-w988d" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070823 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4b264ec9-951e-4928-b43c-5c045b7681f9-cni-binary-copy\") pod \"multus-additional-cni-plugins-hmwzh\" (UID: \"4b264ec9-951e-4928-b43c-5c045b7681f9\") " pod="openshift-multus/multus-additional-cni-plugins-hmwzh" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070843 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6fs6\" (UniqueName: \"kubernetes.io/projected/da1d702d-bbe7-4e02-aa18-cb6556383674-kube-api-access-p6fs6\") pod \"node-resolver-w988d\" (UID: \"da1d702d-bbe7-4e02-aa18-cb6556383674\") " pod="openshift-dns/node-resolver-w988d" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070861 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070885 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070904 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/10e8b890-7f20-4a36-8e03-898620cf599a-proxy-tls\") pod \"machine-config-daemon-747gk\" (UID: \"10e8b890-7f20-4a36-8e03-898620cf599a\") " pod="openshift-machine-config-operator/machine-config-daemon-747gk" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070933 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4b264ec9-951e-4928-b43c-5c045b7681f9-system-cni-dir\") pod \"multus-additional-cni-plugins-hmwzh\" (UID: \"4b264ec9-951e-4928-b43c-5c045b7681f9\") " pod="openshift-multus/multus-additional-cni-plugins-hmwzh" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070956 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070977 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-host-var-lib-cni-multus\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071003 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-run-systemd\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071027 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-node-log\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071056 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071089 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/10e8b890-7f20-4a36-8e03-898620cf599a-rootfs\") pod \"machine-config-daemon-747gk\" (UID: \"10e8b890-7f20-4a36-8e03-898620cf599a\") " pod="openshift-machine-config-operator/machine-config-daemon-747gk" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071108 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-host-var-lib-cni-bin\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071127 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-multus-conf-dir\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071146 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2ff48c28-d076-46e8-a93f-9630989f81e8-ovn-node-metrics-cert\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071163 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbvzp\" (UniqueName: \"kubernetes.io/projected/2ff48c28-d076-46e8-a93f-9630989f81e8-kube-api-access-tbvzp\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071184 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071202 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-host-run-multus-certs\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071222 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071244 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071262 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-host-run-netns\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071280 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-kubelet\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071296 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-log-socket\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071310 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2ff48c28-d076-46e8-a93f-9630989f81e8-ovnkube-script-lib\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071326 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4b264ec9-951e-4928-b43c-5c045b7681f9-os-release\") pod \"multus-additional-cni-plugins-hmwzh\" (UID: \"4b264ec9-951e-4928-b43c-5c045b7681f9\") " pod="openshift-multus/multus-additional-cni-plugins-hmwzh" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071342 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071359 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-host-var-lib-kubelet\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071421 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndw2h\" (UniqueName: \"kubernetes.io/projected/cbe233ee-1ea7-433e-a53a-e4a668f739ee-kube-api-access-ndw2h\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071447 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-var-lib-openvswitch\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071468 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/4b264ec9-951e-4928-b43c-5c045b7681f9-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-hmwzh\" (UID: \"4b264ec9-951e-4928-b43c-5c045b7681f9\") " pod="openshift-multus/multus-additional-cni-plugins-hmwzh" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071490 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071507 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-etc-kubernetes\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071523 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-run-netns\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071541 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-run-ovn-kubernetes\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071572 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4b264ec9-951e-4928-b43c-5c045b7681f9-cnibin\") pod \"multus-additional-cni-plugins-hmwzh\" (UID: \"4b264ec9-951e-4928-b43c-5c045b7681f9\") " pod="openshift-multus/multus-additional-cni-plugins-hmwzh" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071589 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6q8j\" (UniqueName: \"kubernetes.io/projected/4b264ec9-951e-4928-b43c-5c045b7681f9-kube-api-access-v6q8j\") pod \"multus-additional-cni-plugins-hmwzh\" (UID: \"4b264ec9-951e-4928-b43c-5c045b7681f9\") " pod="openshift-multus/multus-additional-cni-plugins-hmwzh" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071608 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071629 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/cbe233ee-1ea7-433e-a53a-e4a668f739ee-multus-daemon-config\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071645 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-slash\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071663 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zv2rl\" (UniqueName: \"kubernetes.io/projected/10e8b890-7f20-4a36-8e03-898620cf599a-kube-api-access-zv2rl\") pod \"machine-config-daemon-747gk\" (UID: \"10e8b890-7f20-4a36-8e03-898620cf599a\") " pod="openshift-machine-config-operator/machine-config-daemon-747gk" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071682 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/cbe233ee-1ea7-433e-a53a-e4a668f739ee-cni-binary-copy\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071700 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071720 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071736 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-host-run-k8s-cni-cncf-io\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071753 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-run-openvswitch\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071769 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-cni-netd\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071790 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/10e8b890-7f20-4a36-8e03-898620cf599a-mcd-auth-proxy-config\") pod \"machine-config-daemon-747gk\" (UID: \"10e8b890-7f20-4a36-8e03-898620cf599a\") " pod="openshift-machine-config-operator/machine-config-daemon-747gk" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071813 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-multus-cni-dir\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071833 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071850 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071868 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-hostroot\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071884 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-cnibin\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071899 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-multus-socket-dir-parent\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071918 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-systemd-units\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071933 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-etc-openvswitch\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071953 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-system-cni-dir\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071967 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-cni-bin\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071983 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4b264ec9-951e-4928-b43c-5c045b7681f9-tuning-conf-dir\") pod \"multus-additional-cni-plugins-hmwzh\" (UID: \"4b264ec9-951e-4928-b43c-5c045b7681f9\") " pod="openshift-multus/multus-additional-cni-plugins-hmwzh" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.072000 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-os-release\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.072016 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-run-ovn\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.072035 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2ff48c28-d076-46e8-a93f-9630989f81e8-ovnkube-config\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.072052 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2ff48c28-d076-46e8-a93f-9630989f81e8-env-overrides\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.072070 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.072144 4840 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.072160 4840 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.072173 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.072185 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.072198 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.072211 4840 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.072223 4840 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.072234 4840 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.072244 4840 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.072256 4840 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.072266 4840 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.072277 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.072286 4840 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.072297 4840 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.072676 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.072700 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.072711 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.072728 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.072739 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:39Z","lastTransitionTime":"2025-09-30T13:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.079197 4840 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.082891 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.089457 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.089599 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069201 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069355 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.068846 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.069372 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070400 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070498 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070726 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070800 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070894 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.070917 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071011 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071410 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071647 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071654 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.104329 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071747 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071766 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071788 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071785 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071803 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071443 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.071919 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.072026 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.072139 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.072208 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.072244 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.072282 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.072258 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.072529 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.072789 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.072799 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.072896 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.073078 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.073062 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.073211 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.073231 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.073253 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.073307 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.073483 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.073581 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.073756 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.102435 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.107468 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.107511 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.105641 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.103199 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.073904 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.073947 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.073968 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.074163 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.074265 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.074396 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.074527 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.075240 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.075444 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.075848 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.075933 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.076371 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.076650 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.077023 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.077242 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.077252 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.077504 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.077616 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.077771 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.078157 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.078340 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.078939 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: E0930 13:56:39.079177 4840 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.079363 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.079404 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.079509 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.079583 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.079612 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.079762 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.079875 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.079956 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.080160 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.080316 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.080515 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.080783 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.080930 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.081220 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.081444 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.081488 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.082071 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.082096 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.082104 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.082640 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.082693 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.083053 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.083859 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.084400 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.084587 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.085037 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.085389 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.085587 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.085613 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.085922 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.085992 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.086014 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.086027 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.086062 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.086072 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.086094 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.084742 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.086763 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.087213 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.087466 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.087309 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.087477 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.087536 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.087709 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.087892 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.088058 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.088237 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.088517 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.088810 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.088860 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.089012 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.089176 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.089703 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.089920 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.091205 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.091221 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.091302 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.091608 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.091987 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.092585 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.092755 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.092970 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.093349 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.093701 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.094228 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.094664 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.094901 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.095287 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.095302 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.095326 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.095356 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.095443 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.096742 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.097925 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.098011 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.098030 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.100323 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.100405 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.100708 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.100808 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.101071 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.101229 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.101184 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.101766 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.101782 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.101986 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.102158 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.102501 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: E0930 13:56:39.102679 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:56:39.602643948 +0000 UTC m=+28.231730371 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.102680 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: E0930 13:56:39.102987 4840 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.103004 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.103107 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.103822 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.103638 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.104688 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.104773 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.104847 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.104831 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: E0930 13:56:39.104956 4840 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.105051 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.105341 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.105724 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.105735 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.105775 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.106022 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.106031 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: E0930 13:56:39.106144 4840 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.106417 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.106976 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.107847 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.107887 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: E0930 13:56:39.109057 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 13:56:39.609031367 +0000 UTC m=+28.238117790 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 13:56:39 crc kubenswrapper[4840]: E0930 13:56:39.109318 4840 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 13:56:39 crc kubenswrapper[4840]: E0930 13:56:39.109336 4840 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 13:56:39 crc kubenswrapper[4840]: E0930 13:56:39.109412 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-30 13:56:39.609402097 +0000 UTC m=+28.238488520 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 13:56:39 crc kubenswrapper[4840]: E0930 13:56:39.109485 4840 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 13:56:39 crc kubenswrapper[4840]: E0930 13:56:39.109498 4840 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 13:56:39 crc kubenswrapper[4840]: E0930 13:56:39.109530 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-30 13:56:39.60951866 +0000 UTC m=+28.238605213 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 13:56:39 crc kubenswrapper[4840]: E0930 13:56:39.109631 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 13:56:39.609613022 +0000 UTC m=+28.238700155 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.112667 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.113737 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.114941 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.115734 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.117349 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.120365 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.123708 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.125215 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.129339 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.132280 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.136599 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.143291 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.147043 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.153048 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 13:56:39 crc kubenswrapper[4840]: W0930 13:56:39.159687 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-530725aac8013eb72bd4d31adcd9b6a122d21084b5071812c3c1a468a500bff0 WatchSource:0}: Error finding container 530725aac8013eb72bd4d31adcd9b6a122d21084b5071812c3c1a468a500bff0: Status 404 returned error can't find the container with id 530725aac8013eb72bd4d31adcd9b6a122d21084b5071812c3c1a468a500bff0 Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.164414 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.173868 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-cni-bin\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.173990 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4b264ec9-951e-4928-b43c-5c045b7681f9-tuning-conf-dir\") pod \"multus-additional-cni-plugins-hmwzh\" (UID: \"4b264ec9-951e-4928-b43c-5c045b7681f9\") " pod="openshift-multus/multus-additional-cni-plugins-hmwzh" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.174033 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-cni-bin\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.174196 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-system-cni-dir\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.174491 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2ff48c28-d076-46e8-a93f-9630989f81e8-ovnkube-config\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.174606 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2ff48c28-d076-46e8-a93f-9630989f81e8-env-overrides\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.174731 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-os-release\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.174858 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-run-ovn\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.174956 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-os-release\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.174295 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-system-cni-dir\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.175089 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/da1d702d-bbe7-4e02-aa18-cb6556383674-hosts-file\") pod \"node-resolver-w988d\" (UID: \"da1d702d-bbe7-4e02-aa18-cb6556383674\") " pod="openshift-dns/node-resolver-w988d" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.175175 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4b264ec9-951e-4928-b43c-5c045b7681f9-cni-binary-copy\") pod \"multus-additional-cni-plugins-hmwzh\" (UID: \"4b264ec9-951e-4928-b43c-5c045b7681f9\") " pod="openshift-multus/multus-additional-cni-plugins-hmwzh" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.175260 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/10e8b890-7f20-4a36-8e03-898620cf599a-proxy-tls\") pod \"machine-config-daemon-747gk\" (UID: \"10e8b890-7f20-4a36-8e03-898620cf599a\") " pod="openshift-machine-config-operator/machine-config-daemon-747gk" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.175146 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/da1d702d-bbe7-4e02-aa18-cb6556383674-hosts-file\") pod \"node-resolver-w988d\" (UID: \"da1d702d-bbe7-4e02-aa18-cb6556383674\") " pod="openshift-dns/node-resolver-w988d" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.175368 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2ff48c28-d076-46e8-a93f-9630989f81e8-env-overrides\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.175119 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-run-ovn\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.175455 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4b264ec9-951e-4928-b43c-5c045b7681f9-system-cni-dir\") pod \"multus-additional-cni-plugins-hmwzh\" (UID: \"4b264ec9-951e-4928-b43c-5c045b7681f9\") " pod="openshift-multus/multus-additional-cni-plugins-hmwzh" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.175533 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6fs6\" (UniqueName: \"kubernetes.io/projected/da1d702d-bbe7-4e02-aa18-cb6556383674-kube-api-access-p6fs6\") pod \"node-resolver-w988d\" (UID: \"da1d702d-bbe7-4e02-aa18-cb6556383674\") " pod="openshift-dns/node-resolver-w988d" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.175635 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.175700 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.175496 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2ff48c28-d076-46e8-a93f-9630989f81e8-ovnkube-config\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.175589 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.175575 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4b264ec9-951e-4928-b43c-5c045b7681f9-system-cni-dir\") pod \"multus-additional-cni-plugins-hmwzh\" (UID: \"4b264ec9-951e-4928-b43c-5c045b7681f9\") " pod="openshift-multus/multus-additional-cni-plugins-hmwzh" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.175973 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/10e8b890-7f20-4a36-8e03-898620cf599a-rootfs\") pod \"machine-config-daemon-747gk\" (UID: \"10e8b890-7f20-4a36-8e03-898620cf599a\") " pod="openshift-machine-config-operator/machine-config-daemon-747gk" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.176031 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/10e8b890-7f20-4a36-8e03-898620cf599a-rootfs\") pod \"machine-config-daemon-747gk\" (UID: \"10e8b890-7f20-4a36-8e03-898620cf599a\") " pod="openshift-machine-config-operator/machine-config-daemon-747gk" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.175996 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.176172 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.176188 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.176209 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.176223 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:39Z","lastTransitionTime":"2025-09-30T13:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.176383 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4b264ec9-951e-4928-b43c-5c045b7681f9-cni-binary-copy\") pod \"multus-additional-cni-plugins-hmwzh\" (UID: \"4b264ec9-951e-4928-b43c-5c045b7681f9\") " pod="openshift-multus/multus-additional-cni-plugins-hmwzh" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.176116 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-host-var-lib-cni-multus\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.176453 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-run-systemd\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.176471 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-node-log\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.176490 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2ff48c28-d076-46e8-a93f-9630989f81e8-ovn-node-metrics-cert\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.176506 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbvzp\" (UniqueName: \"kubernetes.io/projected/2ff48c28-d076-46e8-a93f-9630989f81e8-kube-api-access-tbvzp\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.176601 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-host-var-lib-cni-bin\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.176616 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-multus-conf-dir\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.176709 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-host-var-lib-cni-multus\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.176813 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-host-run-multus-certs\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.176947 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2ff48c28-d076-46e8-a93f-9630989f81e8-ovnkube-script-lib\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.177067 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4b264ec9-951e-4928-b43c-5c045b7681f9-os-release\") pod \"multus-additional-cni-plugins-hmwzh\" (UID: \"4b264ec9-951e-4928-b43c-5c045b7681f9\") " pod="openshift-multus/multus-additional-cni-plugins-hmwzh" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.176886 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-multus-conf-dir\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.176917 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-node-log\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.177192 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4b264ec9-951e-4928-b43c-5c045b7681f9-os-release\") pod \"multus-additional-cni-plugins-hmwzh\" (UID: \"4b264ec9-951e-4928-b43c-5c045b7681f9\") " pod="openshift-multus/multus-additional-cni-plugins-hmwzh" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.176853 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-host-var-lib-cni-bin\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.177161 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-host-run-netns\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.176866 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-run-systemd\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.177292 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-kubelet\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.177326 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-log-socket\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.177353 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-var-lib-openvswitch\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.177354 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-kubelet\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.177377 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/4b264ec9-951e-4928-b43c-5c045b7681f9-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-hmwzh\" (UID: \"4b264ec9-951e-4928-b43c-5c045b7681f9\") " pod="openshift-multus/multus-additional-cni-plugins-hmwzh" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.177382 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-log-socket\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.177367 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-host-run-multus-certs\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.177432 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.177409 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-var-lib-openvswitch\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.177465 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-host-var-lib-kubelet\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.177488 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndw2h\" (UniqueName: \"kubernetes.io/projected/cbe233ee-1ea7-433e-a53a-e4a668f739ee-kube-api-access-ndw2h\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.177510 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-run-netns\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.177529 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-run-ovn-kubernetes\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.177588 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2ff48c28-d076-46e8-a93f-9630989f81e8-ovnkube-script-lib\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.177601 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4b264ec9-951e-4928-b43c-5c045b7681f9-cnibin\") pod \"multus-additional-cni-plugins-hmwzh\" (UID: \"4b264ec9-951e-4928-b43c-5c045b7681f9\") " pod="openshift-multus/multus-additional-cni-plugins-hmwzh" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.177633 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4b264ec9-951e-4928-b43c-5c045b7681f9-cnibin\") pod \"multus-additional-cni-plugins-hmwzh\" (UID: \"4b264ec9-951e-4928-b43c-5c045b7681f9\") " pod="openshift-multus/multus-additional-cni-plugins-hmwzh" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.177645 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-etc-kubernetes\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.177663 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/cbe233ee-1ea7-433e-a53a-e4a668f739ee-multus-daemon-config\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.177673 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-run-netns\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.177679 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-slash\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.177700 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-slash\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.177717 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zv2rl\" (UniqueName: \"kubernetes.io/projected/10e8b890-7f20-4a36-8e03-898620cf599a-kube-api-access-zv2rl\") pod \"machine-config-daemon-747gk\" (UID: \"10e8b890-7f20-4a36-8e03-898620cf599a\") " pod="openshift-machine-config-operator/machine-config-daemon-747gk" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.177727 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-etc-kubernetes\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.177783 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6q8j\" (UniqueName: \"kubernetes.io/projected/4b264ec9-951e-4928-b43c-5c045b7681f9-kube-api-access-v6q8j\") pod \"multus-additional-cni-plugins-hmwzh\" (UID: \"4b264ec9-951e-4928-b43c-5c045b7681f9\") " pod="openshift-multus/multus-additional-cni-plugins-hmwzh" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.177863 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/cbe233ee-1ea7-433e-a53a-e4a668f739ee-cni-binary-copy\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.177969 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-cni-netd\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.177994 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/10e8b890-7f20-4a36-8e03-898620cf599a-mcd-auth-proxy-config\") pod \"machine-config-daemon-747gk\" (UID: \"10e8b890-7f20-4a36-8e03-898620cf599a\") " pod="openshift-machine-config-operator/machine-config-daemon-747gk" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.177874 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-host-var-lib-kubelet\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.178018 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-host-run-k8s-cni-cncf-io\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.178067 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-run-openvswitch\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.178090 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.178147 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-run-ovn-kubernetes\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.178292 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.178404 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-cni-netd\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.178421 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-host-run-k8s-cni-cncf-io\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.178443 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/cbe233ee-1ea7-433e-a53a-e4a668f739ee-multus-daemon-config\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.178440 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-run-openvswitch\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.178536 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.178761 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/cbe233ee-1ea7-433e-a53a-e4a668f739ee-cni-binary-copy\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.178833 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-multus-cni-dir\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.178884 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-hostroot\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.178904 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-multus-cni-dir\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.178907 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-systemd-units\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.178933 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-systemd-units\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.178955 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-etc-openvswitch\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.178983 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-cnibin\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.179005 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-multus-socket-dir-parent\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.179085 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-etc-openvswitch\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.179008 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-hostroot\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.179156 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-multus-socket-dir-parent\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.179195 4840 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.179206 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-cnibin\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.179213 4840 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.179420 4840 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.179440 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.179455 4840 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.179468 4840 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.179487 4840 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.179500 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.179512 4840 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.179524 4840 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.179535 4840 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.179547 4840 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.179578 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.179606 4840 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.179622 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.179634 4840 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.179646 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.179659 4840 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.179672 4840 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.177919 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/4b264ec9-951e-4928-b43c-5c045b7681f9-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-hmwzh\" (UID: \"4b264ec9-951e-4928-b43c-5c045b7681f9\") " pod="openshift-multus/multus-additional-cni-plugins-hmwzh" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.179776 4840 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.179795 4840 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.179810 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.179824 4840 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.179837 4840 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.179698 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/10e8b890-7f20-4a36-8e03-898620cf599a-mcd-auth-proxy-config\") pod \"machine-config-daemon-747gk\" (UID: \"10e8b890-7f20-4a36-8e03-898620cf599a\") " pod="openshift-machine-config-operator/machine-config-daemon-747gk" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.179849 4840 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180063 4840 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180076 4840 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180113 4840 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180123 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180135 4840 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180145 4840 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180156 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180167 4840 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180175 4840 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180184 4840 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180195 4840 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180204 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180214 4840 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180223 4840 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180232 4840 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180243 4840 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180252 4840 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180261 4840 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180270 4840 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180278 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180289 4840 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180300 4840 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180313 4840 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180326 4840 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180337 4840 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180346 4840 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180355 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180365 4840 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180375 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180386 4840 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180398 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180409 4840 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180420 4840 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180431 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180442 4840 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180453 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180464 4840 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180476 4840 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180487 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180498 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180509 4840 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180519 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180529 4840 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180540 4840 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180571 4840 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180583 4840 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180580 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4b264ec9-951e-4928-b43c-5c045b7681f9-tuning-conf-dir\") pod \"multus-additional-cni-plugins-hmwzh\" (UID: \"4b264ec9-951e-4928-b43c-5c045b7681f9\") " pod="openshift-multus/multus-additional-cni-plugins-hmwzh" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180595 4840 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180630 4840 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180644 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180655 4840 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180667 4840 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180678 4840 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180689 4840 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180702 4840 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180714 4840 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.180726 4840 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181337 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181395 4840 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181411 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181423 4840 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181434 4840 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181445 4840 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181456 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181465 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181475 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181483 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181492 4840 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181501 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181510 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181519 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181527 4840 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181535 4840 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181543 4840 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181585 4840 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181595 4840 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181605 4840 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181616 4840 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181625 4840 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181634 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181643 4840 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181652 4840 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181660 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181671 4840 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181679 4840 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181688 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181696 4840 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181705 4840 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181714 4840 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181723 4840 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181732 4840 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181741 4840 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181749 4840 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181759 4840 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181769 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181778 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181788 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181796 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181804 4840 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181813 4840 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181822 4840 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181830 4840 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181838 4840 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181846 4840 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181854 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181862 4840 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181870 4840 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181878 4840 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181887 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181896 4840 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181904 4840 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181912 4840 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181920 4840 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181931 4840 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181942 4840 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181953 4840 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181966 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181974 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181983 4840 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.181991 4840 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.182001 4840 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.182009 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.182022 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.182034 4840 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.182045 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.182059 4840 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.182068 4840 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.182077 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.182085 4840 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.182093 4840 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.182101 4840 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.182111 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.182123 4840 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.182135 4840 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.182149 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.182160 4840 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.182169 4840 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.182177 4840 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.182185 4840 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.182195 4840 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.182205 4840 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.182217 4840 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.182229 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.182241 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.182252 4840 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.182263 4840 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.182271 4840 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.182281 4840 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.182291 4840 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.182302 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.182317 4840 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.182327 4840 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.182341 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.182353 4840 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.182365 4840 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.182378 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.185078 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/10e8b890-7f20-4a36-8e03-898620cf599a-proxy-tls\") pod \"machine-config-daemon-747gk\" (UID: \"10e8b890-7f20-4a36-8e03-898620cf599a\") " pod="openshift-machine-config-operator/machine-config-daemon-747gk" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.185116 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2ff48c28-d076-46e8-a93f-9630989f81e8-ovn-node-metrics-cert\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.185391 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/cbe233ee-1ea7-433e-a53a-e4a668f739ee-host-run-netns\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.186997 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.191600 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbvzp\" (UniqueName: \"kubernetes.io/projected/2ff48c28-d076-46e8-a93f-9630989f81e8-kube-api-access-tbvzp\") pod \"ovnkube-node-kfx69\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.191821 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6fs6\" (UniqueName: \"kubernetes.io/projected/da1d702d-bbe7-4e02-aa18-cb6556383674-kube-api-access-p6fs6\") pod \"node-resolver-w988d\" (UID: \"da1d702d-bbe7-4e02-aa18-cb6556383674\") " pod="openshift-dns/node-resolver-w988d" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.192609 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndw2h\" (UniqueName: \"kubernetes.io/projected/cbe233ee-1ea7-433e-a53a-e4a668f739ee-kube-api-access-ndw2h\") pod \"multus-bwvl2\" (UID: \"cbe233ee-1ea7-433e-a53a-e4a668f739ee\") " pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.194374 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.195903 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zv2rl\" (UniqueName: \"kubernetes.io/projected/10e8b890-7f20-4a36-8e03-898620cf599a-kube-api-access-zv2rl\") pod \"machine-config-daemon-747gk\" (UID: \"10e8b890-7f20-4a36-8e03-898620cf599a\") " pod="openshift-machine-config-operator/machine-config-daemon-747gk" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.196815 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6q8j\" (UniqueName: \"kubernetes.io/projected/4b264ec9-951e-4928-b43c-5c045b7681f9-kube-api-access-v6q8j\") pod \"multus-additional-cni-plugins-hmwzh\" (UID: \"4b264ec9-951e-4928-b43c-5c045b7681f9\") " pod="openshift-multus/multus-additional-cni-plugins-hmwzh" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.270498 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"530725aac8013eb72bd4d31adcd9b6a122d21084b5071812c3c1a468a500bff0"} Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.278891 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.278958 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.278970 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.279043 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.279068 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:39Z","lastTransitionTime":"2025-09-30T13:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.334310 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.344796 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-747gk" Sep 30 13:56:39 crc kubenswrapper[4840]: W0930 13:56:39.345868 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-21725b00569bc97b1ae429d82c34403555478cee483afee4deb5f24f6e3dad13 WatchSource:0}: Error finding container 21725b00569bc97b1ae429d82c34403555478cee483afee4deb5f24f6e3dad13: Status 404 returned error can't find the container with id 21725b00569bc97b1ae429d82c34403555478cee483afee4deb5f24f6e3dad13 Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.353723 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" Sep 30 13:56:39 crc kubenswrapper[4840]: W0930 13:56:39.355842 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10e8b890_7f20_4a36_8e03_898620cf599a.slice/crio-bf1fe0b2b6491572a39a1ccca1af754c2a4db57be70c6fdf4991f41a9585c260 WatchSource:0}: Error finding container bf1fe0b2b6491572a39a1ccca1af754c2a4db57be70c6fdf4991f41a9585c260: Status 404 returned error can't find the container with id bf1fe0b2b6491572a39a1ccca1af754c2a4db57be70c6fdf4991f41a9585c260 Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.365048 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.381294 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.381323 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.381332 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.381348 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.381357 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:39Z","lastTransitionTime":"2025-09-30T13:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:39 crc kubenswrapper[4840]: W0930 13:56:39.381646 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4b264ec9_951e_4928_b43c_5c045b7681f9.slice/crio-a1af42e240a6a2c293ae1eb30355576507ade32370df242e1ec5bab4d939833a WatchSource:0}: Error finding container a1af42e240a6a2c293ae1eb30355576507ade32370df242e1ec5bab4d939833a: Status 404 returned error can't find the container with id a1af42e240a6a2c293ae1eb30355576507ade32370df242e1ec5bab4d939833a Sep 30 13:56:39 crc kubenswrapper[4840]: W0930 13:56:39.386526 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-82a4ce99e70dd7c08b35be95ab7a9171dae028900d044d5b3e2da6fe96b77e81 WatchSource:0}: Error finding container 82a4ce99e70dd7c08b35be95ab7a9171dae028900d044d5b3e2da6fe96b77e81: Status 404 returned error can't find the container with id 82a4ce99e70dd7c08b35be95ab7a9171dae028900d044d5b3e2da6fe96b77e81 Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.401733 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.428192 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-w988d" Sep 30 13:56:39 crc kubenswrapper[4840]: W0930 13:56:39.428293 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2ff48c28_d076_46e8_a93f_9630989f81e8.slice/crio-a9718c1e08d1bd053584576c34795f618d2cd2f04bffabc0aeb0c96f577efab8 WatchSource:0}: Error finding container a9718c1e08d1bd053584576c34795f618d2cd2f04bffabc0aeb0c96f577efab8: Status 404 returned error can't find the container with id a9718c1e08d1bd053584576c34795f618d2cd2f04bffabc0aeb0c96f577efab8 Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.442778 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-bwvl2" Sep 30 13:56:39 crc kubenswrapper[4840]: W0930 13:56:39.445624 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podda1d702d_bbe7_4e02_aa18_cb6556383674.slice/crio-7e80855914b64fc9ac0a32af3b72b1cb7eb1676393cf142cf24c9a9d8a96abc7 WatchSource:0}: Error finding container 7e80855914b64fc9ac0a32af3b72b1cb7eb1676393cf142cf24c9a9d8a96abc7: Status 404 returned error can't find the container with id 7e80855914b64fc9ac0a32af3b72b1cb7eb1676393cf142cf24c9a9d8a96abc7 Sep 30 13:56:39 crc kubenswrapper[4840]: W0930 13:56:39.458157 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcbe233ee_1ea7_433e_a53a_e4a668f739ee.slice/crio-8548a77fd237d1d60349c3f43d9b4ee156a8b18eadd1bcc3596d6f2cbde55c35 WatchSource:0}: Error finding container 8548a77fd237d1d60349c3f43d9b4ee156a8b18eadd1bcc3596d6f2cbde55c35: Status 404 returned error can't find the container with id 8548a77fd237d1d60349c3f43d9b4ee156a8b18eadd1bcc3596d6f2cbde55c35 Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.484471 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.484507 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.484541 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.484574 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.484585 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:39Z","lastTransitionTime":"2025-09-30T13:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.587162 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.587204 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.587215 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.587234 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.587245 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:39Z","lastTransitionTime":"2025-09-30T13:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.686437 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.686630 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:56:39 crc kubenswrapper[4840]: E0930 13:56:39.686697 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:56:40.686660049 +0000 UTC m=+29.315746472 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.686759 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:56:39 crc kubenswrapper[4840]: E0930 13:56:39.686780 4840 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 13:56:39 crc kubenswrapper[4840]: E0930 13:56:39.686856 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 13:56:40.686835054 +0000 UTC m=+29.315921477 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.686884 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.686921 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:56:39 crc kubenswrapper[4840]: E0930 13:56:39.686928 4840 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 13:56:39 crc kubenswrapper[4840]: E0930 13:56:39.686970 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 13:56:40.686963627 +0000 UTC m=+29.316050050 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 13:56:39 crc kubenswrapper[4840]: E0930 13:56:39.687020 4840 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 13:56:39 crc kubenswrapper[4840]: E0930 13:56:39.687037 4840 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 13:56:39 crc kubenswrapper[4840]: E0930 13:56:39.687052 4840 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 13:56:39 crc kubenswrapper[4840]: E0930 13:56:39.687062 4840 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 13:56:39 crc kubenswrapper[4840]: E0930 13:56:39.687074 4840 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 13:56:39 crc kubenswrapper[4840]: E0930 13:56:39.687087 4840 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 13:56:39 crc kubenswrapper[4840]: E0930 13:56:39.687087 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-30 13:56:40.68707906 +0000 UTC m=+29.316165733 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 13:56:39 crc kubenswrapper[4840]: E0930 13:56:39.687216 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-30 13:56:40.687155862 +0000 UTC m=+29.316242285 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.689367 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.689423 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.689436 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.689460 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.689476 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:39Z","lastTransitionTime":"2025-09-30T13:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.791710 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.792039 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.792050 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.792074 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.792086 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:39Z","lastTransitionTime":"2025-09-30T13:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.894876 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.894946 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.894963 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.894992 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.895011 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:39Z","lastTransitionTime":"2025-09-30T13:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.998301 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.998347 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.998356 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.998375 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:39 crc kubenswrapper[4840]: I0930 13:56:39.998388 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:39Z","lastTransitionTime":"2025-09-30T13:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.101258 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.101335 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.101362 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.101399 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.101425 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:40Z","lastTransitionTime":"2025-09-30T13:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.115816 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:56:40 crc kubenswrapper[4840]: E0930 13:56:40.115991 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.120412 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.121428 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.123024 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.124289 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.125169 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.125910 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.126665 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.127238 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.127937 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.128488 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.129144 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.129935 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.131290 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.132497 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.133496 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.134360 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.135318 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.138085 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.140636 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.143440 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.144453 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.145340 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.146025 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.146995 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.147586 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.148387 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.149318 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.150008 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.150822 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.151466 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.152139 4840 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.152278 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.155153 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.155880 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.156428 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.157707 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.158481 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.159065 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.159757 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.160525 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.161107 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.161765 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.162374 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.163047 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.164460 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.165215 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.166631 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.167497 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.168534 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.169220 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.170150 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.170795 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.171498 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.172719 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.205122 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.205181 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.205194 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.205217 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.205231 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:40Z","lastTransitionTime":"2025-09-30T13:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.274604 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" event={"ID":"10e8b890-7f20-4a36-8e03-898620cf599a","Type":"ContainerStarted","Data":"91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13"} Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.274662 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" event={"ID":"10e8b890-7f20-4a36-8e03-898620cf599a","Type":"ContainerStarted","Data":"ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd"} Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.274677 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" event={"ID":"10e8b890-7f20-4a36-8e03-898620cf599a","Type":"ContainerStarted","Data":"bf1fe0b2b6491572a39a1ccca1af754c2a4db57be70c6fdf4991f41a9585c260"} Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.275879 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-w988d" event={"ID":"da1d702d-bbe7-4e02-aa18-cb6556383674","Type":"ContainerStarted","Data":"4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d"} Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.275906 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-w988d" event={"ID":"da1d702d-bbe7-4e02-aa18-cb6556383674","Type":"ContainerStarted","Data":"7e80855914b64fc9ac0a32af3b72b1cb7eb1676393cf142cf24c9a9d8a96abc7"} Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.277385 4840 generic.go:334] "Generic (PLEG): container finished" podID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerID="c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30" exitCode=0 Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.277436 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" event={"ID":"2ff48c28-d076-46e8-a93f-9630989f81e8","Type":"ContainerDied","Data":"c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30"} Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.277455 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" event={"ID":"2ff48c28-d076-46e8-a93f-9630989f81e8","Type":"ContainerStarted","Data":"a9718c1e08d1bd053584576c34795f618d2cd2f04bffabc0aeb0c96f577efab8"} Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.279448 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"82a4ce99e70dd7c08b35be95ab7a9171dae028900d044d5b3e2da6fe96b77e81"} Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.280839 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" event={"ID":"4b264ec9-951e-4928-b43c-5c045b7681f9","Type":"ContainerStarted","Data":"642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8"} Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.280880 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" event={"ID":"4b264ec9-951e-4928-b43c-5c045b7681f9","Type":"ContainerStarted","Data":"a1af42e240a6a2c293ae1eb30355576507ade32370df242e1ec5bab4d939833a"} Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.282907 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b"} Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.282957 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee"} Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.284257 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bwvl2" event={"ID":"cbe233ee-1ea7-433e-a53a-e4a668f739ee","Type":"ContainerStarted","Data":"1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a"} Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.284292 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bwvl2" event={"ID":"cbe233ee-1ea7-433e-a53a-e4a668f739ee","Type":"ContainerStarted","Data":"8548a77fd237d1d60349c3f43d9b4ee156a8b18eadd1bcc3596d6f2cbde55c35"} Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.285563 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee"} Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.285595 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"21725b00569bc97b1ae429d82c34403555478cee483afee4deb5f24f6e3dad13"} Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.286709 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.296729 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.307775 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.308045 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.308142 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.308232 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.308330 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:40Z","lastTransitionTime":"2025-09-30T13:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.309057 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.319813 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.328978 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.338989 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.352535 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.365519 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.373221 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.410243 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.410380 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.410486 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.410630 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.410737 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:40Z","lastTransitionTime":"2025-09-30T13:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.410367 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.419589 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.433027 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.444166 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.457203 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.469079 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.478980 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.495339 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.510600 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.513453 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.513499 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.513509 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.513529 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.513539 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:40Z","lastTransitionTime":"2025-09-30T13:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.520870 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.528392 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.536015 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.550530 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.616053 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.616107 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.616122 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.616145 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.616159 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:40Z","lastTransitionTime":"2025-09-30T13:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.697498 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.697661 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:56:40 crc kubenswrapper[4840]: E0930 13:56:40.697936 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:56:42.697891831 +0000 UTC m=+31.326978284 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:56:40 crc kubenswrapper[4840]: E0930 13:56:40.697989 4840 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 13:56:40 crc kubenswrapper[4840]: E0930 13:56:40.698065 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 13:56:42.698047375 +0000 UTC m=+31.327133798 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 13:56:40 crc kubenswrapper[4840]: E0930 13:56:40.698136 4840 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 13:56:40 crc kubenswrapper[4840]: E0930 13:56:40.698162 4840 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 13:56:40 crc kubenswrapper[4840]: E0930 13:56:40.698181 4840 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 13:56:40 crc kubenswrapper[4840]: E0930 13:56:40.698236 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-30 13:56:42.698218639 +0000 UTC m=+31.327305222 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.697993 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.698315 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.698365 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:56:40 crc kubenswrapper[4840]: E0930 13:56:40.698464 4840 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 13:56:40 crc kubenswrapper[4840]: E0930 13:56:40.698522 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 13:56:42.698508477 +0000 UTC m=+31.327594920 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 13:56:40 crc kubenswrapper[4840]: E0930 13:56:40.698629 4840 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 13:56:40 crc kubenswrapper[4840]: E0930 13:56:40.698663 4840 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 13:56:40 crc kubenswrapper[4840]: E0930 13:56:40.698675 4840 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 13:56:40 crc kubenswrapper[4840]: E0930 13:56:40.698733 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-30 13:56:42.698714772 +0000 UTC m=+31.327801365 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.718071 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.718116 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.718125 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.718143 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.718152 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:40Z","lastTransitionTime":"2025-09-30T13:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.820404 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.820437 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.820472 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.820489 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.820499 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:40Z","lastTransitionTime":"2025-09-30T13:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.922386 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.922421 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.922432 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.922461 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:40 crc kubenswrapper[4840]: I0930 13:56:40.922475 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:40Z","lastTransitionTime":"2025-09-30T13:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.024473 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.024529 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.024577 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.024597 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.024608 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:41Z","lastTransitionTime":"2025-09-30T13:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.116290 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.116336 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:56:41 crc kubenswrapper[4840]: E0930 13:56:41.116491 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:56:41 crc kubenswrapper[4840]: E0930 13:56:41.116638 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.127360 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.127393 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.127405 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.127423 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.127434 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:41Z","lastTransitionTime":"2025-09-30T13:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.231042 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.231087 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.231098 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.231118 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.231131 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:41Z","lastTransitionTime":"2025-09-30T13:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.256610 4840 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.272857 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:41Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.275609 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.276989 4840 scope.go:117] "RemoveContainer" containerID="fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb" Sep 30 13:56:41 crc kubenswrapper[4840]: E0930 13:56:41.277157 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.291068 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" event={"ID":"2ff48c28-d076-46e8-a93f-9630989f81e8","Type":"ContainerStarted","Data":"f27a94b18137fa991459484406b1c5bd2b2dbb52f527d5f13156ba8e5b8d0146"} Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.291106 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" event={"ID":"2ff48c28-d076-46e8-a93f-9630989f81e8","Type":"ContainerStarted","Data":"f84052a424fd6279d70b54902ebdcf430c8aa53b02406b52fc261db9d4d88ad5"} Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.291118 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" event={"ID":"2ff48c28-d076-46e8-a93f-9630989f81e8","Type":"ContainerStarted","Data":"b9549ce85242f1f19b3a52e603caec58bb4c3e2449eb50d8c4525f56dfc93e98"} Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.291126 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" event={"ID":"2ff48c28-d076-46e8-a93f-9630989f81e8","Type":"ContainerStarted","Data":"7a364a13c773fe00ee4e022a286821013d3a681acc16567e66b1c33c7e18ee3a"} Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.292300 4840 generic.go:334] "Generic (PLEG): container finished" podID="4b264ec9-951e-4928-b43c-5c045b7681f9" containerID="642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8" exitCode=0 Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.292469 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" event={"ID":"4b264ec9-951e-4928-b43c-5c045b7681f9","Type":"ContainerDied","Data":"642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8"} Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.293320 4840 scope.go:117] "RemoveContainer" containerID="fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb" Sep 30 13:56:41 crc kubenswrapper[4840]: E0930 13:56:41.293471 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.297481 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:41Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.308854 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:41Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.322976 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:41Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.334277 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.334321 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.334334 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.334353 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.334366 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:41Z","lastTransitionTime":"2025-09-30T13:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.340884 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:41Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.354204 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:41Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.366675 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:41Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.379231 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:41Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.395656 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:41Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.407632 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:41Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.418466 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:41Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.428948 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:41Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.437439 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.437485 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.437496 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.437516 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.437528 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:41Z","lastTransitionTime":"2025-09-30T13:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.440040 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:41Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.450691 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:41Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.462825 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:41Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.474968 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:41Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.488417 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:41Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.499595 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:41Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.510675 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:41Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.525565 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f8cd5bd-4a7f-4a5b-ab00-0cc64de9de94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cba320f443ed90a4e755228b663a1da3096e7ea8e068671d1f6273544b6ef60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c507106f39ae2e8acd4268c4e8f28ffd88a0965e69544d8d8a8830e83eceaf9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d62fbefb67a24fedc5f13652596c76f5b288fec58d79b5209f49c79db139d79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\"\\\\nI0930 13:56:34.911386 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 13:56:34.920195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 13:56:34.920220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 13:56:34.920240 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 13:56:34.920248 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 13:56:34.927953 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 13:56:34.927977 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927984 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 13:56:34.927992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 13:56:34.927996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 13:56:34.928001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 13:56:34.928159 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0930 13:56:34.930945 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759240579\\\\\\\\\\\\\\\" (2025-09-30 13:56:19 +0000 UTC to 2025-10-30 13:56:20 +0000 UTC (now=2025-09-30 13:56:34.930900435 +0000 UTC))\\\\\\\"\\\\nF0930 13:56:34.930995 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1435e2e88d45e5561ff5863095d0f953ef1478d19d87691ab226cad7ae148f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:41Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.538459 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:41Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.539541 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.539591 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.539602 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.539618 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.539627 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:41Z","lastTransitionTime":"2025-09-30T13:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.556765 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:41Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.570426 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:41Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.641266 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.641309 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.641321 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.641343 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.641354 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:41Z","lastTransitionTime":"2025-09-30T13:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.743733 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.743776 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.743785 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.743801 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.743810 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:41Z","lastTransitionTime":"2025-09-30T13:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.846822 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.847108 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.847195 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.847282 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.847362 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:41Z","lastTransitionTime":"2025-09-30T13:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.949969 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.950029 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.950045 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.950072 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:41 crc kubenswrapper[4840]: I0930 13:56:41.950094 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:41Z","lastTransitionTime":"2025-09-30T13:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.052456 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.052490 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.052500 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.052516 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.052525 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:42Z","lastTransitionTime":"2025-09-30T13:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.115900 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:56:42 crc kubenswrapper[4840]: E0930 13:56:42.116038 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.135931 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.151878 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.154898 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.154940 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.154955 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.154978 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.154992 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:42Z","lastTransitionTime":"2025-09-30T13:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.169213 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.188742 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.208444 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.224296 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f8cd5bd-4a7f-4a5b-ab00-0cc64de9de94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cba320f443ed90a4e755228b663a1da3096e7ea8e068671d1f6273544b6ef60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c507106f39ae2e8acd4268c4e8f28ffd88a0965e69544d8d8a8830e83eceaf9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d62fbefb67a24fedc5f13652596c76f5b288fec58d79b5209f49c79db139d79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\"\\\\nI0930 13:56:34.911386 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 13:56:34.920195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 13:56:34.920220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 13:56:34.920240 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 13:56:34.920248 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 13:56:34.927953 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 13:56:34.927977 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927984 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 13:56:34.927992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 13:56:34.927996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 13:56:34.928001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 13:56:34.928159 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0930 13:56:34.930945 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759240579\\\\\\\\\\\\\\\" (2025-09-30 13:56:19 +0000 UTC to 2025-10-30 13:56:20 +0000 UTC (now=2025-09-30 13:56:34.930900435 +0000 UTC))\\\\\\\"\\\\nF0930 13:56:34.930995 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1435e2e88d45e5561ff5863095d0f953ef1478d19d87691ab226cad7ae148f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.243138 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.257652 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.257691 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.257704 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.257724 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.257738 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:42Z","lastTransitionTime":"2025-09-30T13:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.262049 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.274392 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.290593 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.297396 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" event={"ID":"4b264ec9-951e-4928-b43c-5c045b7681f9","Type":"ContainerStarted","Data":"d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb"} Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.306200 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" event={"ID":"2ff48c28-d076-46e8-a93f-9630989f81e8","Type":"ContainerStarted","Data":"d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134"} Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.306255 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" event={"ID":"2ff48c28-d076-46e8-a93f-9630989f81e8","Type":"ContainerStarted","Data":"bc4e5d3625aa2d5f94307cedddf72cc48f17f327a044501144e35422d670d555"} Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.316585 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.331224 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.360609 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.360645 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.360655 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.360680 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.360690 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:42Z","lastTransitionTime":"2025-09-30T13:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.369801 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.383911 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.399269 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.416875 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.428950 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.443400 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.457197 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.462908 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.462935 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.462948 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.462966 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.462978 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:42Z","lastTransitionTime":"2025-09-30T13:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.478374 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f8cd5bd-4a7f-4a5b-ab00-0cc64de9de94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cba320f443ed90a4e755228b663a1da3096e7ea8e068671d1f6273544b6ef60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c507106f39ae2e8acd4268c4e8f28ffd88a0965e69544d8d8a8830e83eceaf9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d62fbefb67a24fedc5f13652596c76f5b288fec58d79b5209f49c79db139d79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\"\\\\nI0930 13:56:34.911386 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 13:56:34.920195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 13:56:34.920220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 13:56:34.920240 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 13:56:34.920248 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 13:56:34.927953 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 13:56:34.927977 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927984 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 13:56:34.927992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 13:56:34.927996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 13:56:34.928001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 13:56:34.928159 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0930 13:56:34.930945 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759240579\\\\\\\\\\\\\\\" (2025-09-30 13:56:19 +0000 UTC to 2025-10-30 13:56:20 +0000 UTC (now=2025-09-30 13:56:34.930900435 +0000 UTC))\\\\\\\"\\\\nF0930 13:56:34.930995 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1435e2e88d45e5561ff5863095d0f953ef1478d19d87691ab226cad7ae148f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.489390 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.507058 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.522682 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.535397 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.565158 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.565189 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.565198 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.565213 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.565223 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:42Z","lastTransitionTime":"2025-09-30T13:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.667746 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.668090 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.668101 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.668118 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.668128 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:42Z","lastTransitionTime":"2025-09-30T13:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.718432 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.718563 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.718611 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:56:42 crc kubenswrapper[4840]: E0930 13:56:42.718634 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:56:46.718605376 +0000 UTC m=+35.347691799 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.718698 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.718766 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:56:42 crc kubenswrapper[4840]: E0930 13:56:42.718699 4840 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 13:56:42 crc kubenswrapper[4840]: E0930 13:56:42.718900 4840 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 13:56:42 crc kubenswrapper[4840]: E0930 13:56:42.718911 4840 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 13:56:42 crc kubenswrapper[4840]: E0930 13:56:42.718914 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 13:56:46.718896544 +0000 UTC m=+35.347982967 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 13:56:42 crc kubenswrapper[4840]: E0930 13:56:42.718922 4840 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 13:56:42 crc kubenswrapper[4840]: E0930 13:56:42.718955 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-30 13:56:46.718948556 +0000 UTC m=+35.348034969 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 13:56:42 crc kubenswrapper[4840]: E0930 13:56:42.718770 4840 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 13:56:42 crc kubenswrapper[4840]: E0930 13:56:42.718986 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 13:56:46.718980786 +0000 UTC m=+35.348067209 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 13:56:42 crc kubenswrapper[4840]: E0930 13:56:42.718846 4840 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 13:56:42 crc kubenswrapper[4840]: E0930 13:56:42.719004 4840 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 13:56:42 crc kubenswrapper[4840]: E0930 13:56:42.719012 4840 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 13:56:42 crc kubenswrapper[4840]: E0930 13:56:42.719035 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-30 13:56:46.719029188 +0000 UTC m=+35.348115611 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.770836 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.770878 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.770887 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.770903 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.770912 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:42Z","lastTransitionTime":"2025-09-30T13:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.873837 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.873880 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.873889 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.873906 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.873917 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:42Z","lastTransitionTime":"2025-09-30T13:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.977664 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.977722 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.977735 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.977758 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:42 crc kubenswrapper[4840]: I0930 13:56:42.977771 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:42Z","lastTransitionTime":"2025-09-30T13:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.080230 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.080275 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.080287 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.080307 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.080320 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:43Z","lastTransitionTime":"2025-09-30T13:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.115727 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:56:43 crc kubenswrapper[4840]: E0930 13:56:43.115933 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.115749 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:56:43 crc kubenswrapper[4840]: E0930 13:56:43.116153 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.139040 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-pj2zd"] Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.139464 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-pj2zd" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.143807 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.143934 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.143950 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.145800 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.166179 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:43Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.182933 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:43Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.183037 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.183083 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.183094 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.183113 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.183123 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:43Z","lastTransitionTime":"2025-09-30T13:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.193227 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:43Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.206889 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f8cd5bd-4a7f-4a5b-ab00-0cc64de9de94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cba320f443ed90a4e755228b663a1da3096e7ea8e068671d1f6273544b6ef60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c507106f39ae2e8acd4268c4e8f28ffd88a0965e69544d8d8a8830e83eceaf9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d62fbefb67a24fedc5f13652596c76f5b288fec58d79b5209f49c79db139d79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\"\\\\nI0930 13:56:34.911386 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 13:56:34.920195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 13:56:34.920220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 13:56:34.920240 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 13:56:34.920248 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 13:56:34.927953 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 13:56:34.927977 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927984 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 13:56:34.927992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 13:56:34.927996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 13:56:34.928001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 13:56:34.928159 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0930 13:56:34.930945 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759240579\\\\\\\\\\\\\\\" (2025-09-30 13:56:19 +0000 UTC to 2025-10-30 13:56:20 +0000 UTC (now=2025-09-30 13:56:34.930900435 +0000 UTC))\\\\\\\"\\\\nF0930 13:56:34.930995 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1435e2e88d45e5561ff5863095d0f953ef1478d19d87691ab226cad7ae148f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:43Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.218934 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:43Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.223574 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4de996b8-bda8-4142-b5e0-04ed9ae1f327-host\") pod \"node-ca-pj2zd\" (UID: \"4de996b8-bda8-4142-b5e0-04ed9ae1f327\") " pod="openshift-image-registry/node-ca-pj2zd" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.223653 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/4de996b8-bda8-4142-b5e0-04ed9ae1f327-serviceca\") pod \"node-ca-pj2zd\" (UID: \"4de996b8-bda8-4142-b5e0-04ed9ae1f327\") " pod="openshift-image-registry/node-ca-pj2zd" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.223673 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pn7p6\" (UniqueName: \"kubernetes.io/projected/4de996b8-bda8-4142-b5e0-04ed9ae1f327-kube-api-access-pn7p6\") pod \"node-ca-pj2zd\" (UID: \"4de996b8-bda8-4142-b5e0-04ed9ae1f327\") " pod="openshift-image-registry/node-ca-pj2zd" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.241221 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:43Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.274325 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:43Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.284783 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.284823 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.284832 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.284848 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.284857 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:43Z","lastTransitionTime":"2025-09-30T13:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.298863 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:43Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.310786 4840 generic.go:334] "Generic (PLEG): container finished" podID="4b264ec9-951e-4928-b43c-5c045b7681f9" containerID="d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb" exitCode=0 Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.310872 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" event={"ID":"4b264ec9-951e-4928-b43c-5c045b7681f9","Type":"ContainerDied","Data":"d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb"} Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.312056 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pj2zd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4de996b8-bda8-4142-b5e0-04ed9ae1f327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pj2zd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:43Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.312591 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"2b16edc8f720d4296d7fe50fe407f5983ab975818ec6954d1f0e738e5596624a"} Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.324621 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4de996b8-bda8-4142-b5e0-04ed9ae1f327-host\") pod \"node-ca-pj2zd\" (UID: \"4de996b8-bda8-4142-b5e0-04ed9ae1f327\") " pod="openshift-image-registry/node-ca-pj2zd" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.324707 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/4de996b8-bda8-4142-b5e0-04ed9ae1f327-serviceca\") pod \"node-ca-pj2zd\" (UID: \"4de996b8-bda8-4142-b5e0-04ed9ae1f327\") " pod="openshift-image-registry/node-ca-pj2zd" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.324727 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pn7p6\" (UniqueName: \"kubernetes.io/projected/4de996b8-bda8-4142-b5e0-04ed9ae1f327-kube-api-access-pn7p6\") pod \"node-ca-pj2zd\" (UID: \"4de996b8-bda8-4142-b5e0-04ed9ae1f327\") " pod="openshift-image-registry/node-ca-pj2zd" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.324753 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4de996b8-bda8-4142-b5e0-04ed9ae1f327-host\") pod \"node-ca-pj2zd\" (UID: \"4de996b8-bda8-4142-b5e0-04ed9ae1f327\") " pod="openshift-image-registry/node-ca-pj2zd" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.325517 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/4de996b8-bda8-4142-b5e0-04ed9ae1f327-serviceca\") pod \"node-ca-pj2zd\" (UID: \"4de996b8-bda8-4142-b5e0-04ed9ae1f327\") " pod="openshift-image-registry/node-ca-pj2zd" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.326883 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:43Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.342250 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:43Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.343032 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pn7p6\" (UniqueName: \"kubernetes.io/projected/4de996b8-bda8-4142-b5e0-04ed9ae1f327-kube-api-access-pn7p6\") pod \"node-ca-pj2zd\" (UID: \"4de996b8-bda8-4142-b5e0-04ed9ae1f327\") " pod="openshift-image-registry/node-ca-pj2zd" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.360461 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:43Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.380331 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:43Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.387330 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.387374 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.387384 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.387403 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.387415 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:43Z","lastTransitionTime":"2025-09-30T13:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.395980 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:43Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.412427 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:43Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.428142 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f8cd5bd-4a7f-4a5b-ab00-0cc64de9de94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cba320f443ed90a4e755228b663a1da3096e7ea8e068671d1f6273544b6ef60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c507106f39ae2e8acd4268c4e8f28ffd88a0965e69544d8d8a8830e83eceaf9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d62fbefb67a24fedc5f13652596c76f5b288fec58d79b5209f49c79db139d79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\"\\\\nI0930 13:56:34.911386 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 13:56:34.920195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 13:56:34.920220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 13:56:34.920240 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 13:56:34.920248 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 13:56:34.927953 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 13:56:34.927977 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927984 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 13:56:34.927992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 13:56:34.927996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 13:56:34.928001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 13:56:34.928159 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0930 13:56:34.930945 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759240579\\\\\\\\\\\\\\\" (2025-09-30 13:56:19 +0000 UTC to 2025-10-30 13:56:20 +0000 UTC (now=2025-09-30 13:56:34.930900435 +0000 UTC))\\\\\\\"\\\\nF0930 13:56:34.930995 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1435e2e88d45e5561ff5863095d0f953ef1478d19d87691ab226cad7ae148f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:43Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.439872 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:43Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.455421 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:43Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.455813 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-pj2zd" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.467480 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b16edc8f720d4296d7fe50fe407f5983ab975818ec6954d1f0e738e5596624a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:43Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:43 crc kubenswrapper[4840]: W0930 13:56:43.473117 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4de996b8_bda8_4142_b5e0_04ed9ae1f327.slice/crio-c56c2e57ae95f1e4166f8c0849ae386c76bae8b294dd30a5e084d9afe2ce9cc4 WatchSource:0}: Error finding container c56c2e57ae95f1e4166f8c0849ae386c76bae8b294dd30a5e084d9afe2ce9cc4: Status 404 returned error can't find the container with id c56c2e57ae95f1e4166f8c0849ae386c76bae8b294dd30a5e084d9afe2ce9cc4 Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.479871 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:43Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.490949 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.491001 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.491011 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.491055 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.491065 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:43Z","lastTransitionTime":"2025-09-30T13:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.502387 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:43Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.522117 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:43Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.535256 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:43Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.549423 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:43Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.564218 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:43Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.576378 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pj2zd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4de996b8-bda8-4142-b5e0-04ed9ae1f327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pj2zd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:43Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.593970 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.594021 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.594031 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.594053 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.594064 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:43Z","lastTransitionTime":"2025-09-30T13:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.697110 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.697164 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.697182 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.697212 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.697233 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:43Z","lastTransitionTime":"2025-09-30T13:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.800142 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.800490 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.800504 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.800524 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.800534 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:43Z","lastTransitionTime":"2025-09-30T13:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.902944 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.903025 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.903038 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.903058 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:43 crc kubenswrapper[4840]: I0930 13:56:43.903077 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:43Z","lastTransitionTime":"2025-09-30T13:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.005045 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.005075 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.005086 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.005103 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.005116 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:44Z","lastTransitionTime":"2025-09-30T13:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.109396 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.109440 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.109450 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.109477 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.109691 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:44Z","lastTransitionTime":"2025-09-30T13:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.116638 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:56:44 crc kubenswrapper[4840]: E0930 13:56:44.116793 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.211990 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.212024 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.212033 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.212046 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.212055 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:44Z","lastTransitionTime":"2025-09-30T13:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.314169 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.314212 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.314224 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.314242 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.314253 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:44Z","lastTransitionTime":"2025-09-30T13:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.324753 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" event={"ID":"2ff48c28-d076-46e8-a93f-9630989f81e8","Type":"ContainerStarted","Data":"c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d"} Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.325873 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-pj2zd" event={"ID":"4de996b8-bda8-4142-b5e0-04ed9ae1f327","Type":"ContainerStarted","Data":"7286e2d685d6573993e2fa000cb9e08d85c6aa820af507327be52ca50fe58c7a"} Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.325899 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-pj2zd" event={"ID":"4de996b8-bda8-4142-b5e0-04ed9ae1f327","Type":"ContainerStarted","Data":"c56c2e57ae95f1e4166f8c0849ae386c76bae8b294dd30a5e084d9afe2ce9cc4"} Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.328748 4840 generic.go:334] "Generic (PLEG): container finished" podID="4b264ec9-951e-4928-b43c-5c045b7681f9" containerID="1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228" exitCode=0 Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.328817 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" event={"ID":"4b264ec9-951e-4928-b43c-5c045b7681f9","Type":"ContainerDied","Data":"1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228"} Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.349232 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f8cd5bd-4a7f-4a5b-ab00-0cc64de9de94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cba320f443ed90a4e755228b663a1da3096e7ea8e068671d1f6273544b6ef60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c507106f39ae2e8acd4268c4e8f28ffd88a0965e69544d8d8a8830e83eceaf9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d62fbefb67a24fedc5f13652596c76f5b288fec58d79b5209f49c79db139d79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\"\\\\nI0930 13:56:34.911386 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 13:56:34.920195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 13:56:34.920220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 13:56:34.920240 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 13:56:34.920248 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 13:56:34.927953 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 13:56:34.927977 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927984 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 13:56:34.927992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 13:56:34.927996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 13:56:34.928001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 13:56:34.928159 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0930 13:56:34.930945 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759240579\\\\\\\\\\\\\\\" (2025-09-30 13:56:19 +0000 UTC to 2025-10-30 13:56:20 +0000 UTC (now=2025-09-30 13:56:34.930900435 +0000 UTC))\\\\\\\"\\\\nF0930 13:56:34.930995 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1435e2e88d45e5561ff5863095d0f953ef1478d19d87691ab226cad7ae148f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:44Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.363159 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:44Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.378536 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:44Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.390385 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b16edc8f720d4296d7fe50fe407f5983ab975818ec6954d1f0e738e5596624a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:44Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.401838 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:44Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.417333 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.417381 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.417392 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.417413 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.417426 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:44Z","lastTransitionTime":"2025-09-30T13:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.423155 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:44Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.438683 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:44Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.453835 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:44Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.465767 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:44Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.481061 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:44Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.494018 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pj2zd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4de996b8-bda8-4142-b5e0-04ed9ae1f327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7286e2d685d6573993e2fa000cb9e08d85c6aa820af507327be52ca50fe58c7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pj2zd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:44Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.506590 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:44Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.517719 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:44Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.519423 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.519448 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.519459 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.519477 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.519488 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:44Z","lastTransitionTime":"2025-09-30T13:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.531416 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:44Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.544598 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:44Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.558449 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:44Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.569656 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pj2zd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4de996b8-bda8-4142-b5e0-04ed9ae1f327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7286e2d685d6573993e2fa000cb9e08d85c6aa820af507327be52ca50fe58c7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pj2zd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:44Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.584952 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:44Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.597331 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:44Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.609573 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:44Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.622718 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.622760 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.622774 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.622794 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.622807 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:44Z","lastTransitionTime":"2025-09-30T13:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.635789 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f8cd5bd-4a7f-4a5b-ab00-0cc64de9de94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cba320f443ed90a4e755228b663a1da3096e7ea8e068671d1f6273544b6ef60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c507106f39ae2e8acd4268c4e8f28ffd88a0965e69544d8d8a8830e83eceaf9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d62fbefb67a24fedc5f13652596c76f5b288fec58d79b5209f49c79db139d79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\"\\\\nI0930 13:56:34.911386 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 13:56:34.920195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 13:56:34.920220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 13:56:34.920240 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 13:56:34.920248 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 13:56:34.927953 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 13:56:34.927977 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927984 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 13:56:34.927992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 13:56:34.927996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 13:56:34.928001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 13:56:34.928159 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0930 13:56:34.930945 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759240579\\\\\\\\\\\\\\\" (2025-09-30 13:56:19 +0000 UTC to 2025-10-30 13:56:20 +0000 UTC (now=2025-09-30 13:56:34.930900435 +0000 UTC))\\\\\\\"\\\\nF0930 13:56:34.930995 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1435e2e88d45e5561ff5863095d0f953ef1478d19d87691ab226cad7ae148f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:44Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.650192 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:44Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.665205 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:44Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.677392 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b16edc8f720d4296d7fe50fe407f5983ab975818ec6954d1f0e738e5596624a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:44Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.695850 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:44Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.708355 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:44Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.724908 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.724946 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.724957 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.724975 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.724987 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:44Z","lastTransitionTime":"2025-09-30T13:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.827483 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.827521 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.827532 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.827567 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.827580 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:44Z","lastTransitionTime":"2025-09-30T13:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.930115 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.930198 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.930233 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.930269 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:44 crc kubenswrapper[4840]: I0930 13:56:44.930294 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:44Z","lastTransitionTime":"2025-09-30T13:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.033638 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.033680 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.033692 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.033714 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.033726 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:45Z","lastTransitionTime":"2025-09-30T13:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.115938 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.116022 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:56:45 crc kubenswrapper[4840]: E0930 13:56:45.116132 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:56:45 crc kubenswrapper[4840]: E0930 13:56:45.116220 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.137120 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.137176 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.137190 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.137210 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.137220 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:45Z","lastTransitionTime":"2025-09-30T13:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.240394 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.240429 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.240437 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.240454 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.240463 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:45Z","lastTransitionTime":"2025-09-30T13:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.296113 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.296166 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.296185 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.296201 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.296211 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:45Z","lastTransitionTime":"2025-09-30T13:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:45 crc kubenswrapper[4840]: E0930 13:56:45.309240 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5\\\",\\\"systemUUID\\\":\\\"f41e0351-4eda-411a-9a99-43a1952f3d34\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:45Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.318254 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.318291 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.318346 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.318363 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.318376 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:45Z","lastTransitionTime":"2025-09-30T13:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:45 crc kubenswrapper[4840]: E0930 13:56:45.333836 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5\\\",\\\"systemUUID\\\":\\\"f41e0351-4eda-411a-9a99-43a1952f3d34\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:45Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.335319 4840 generic.go:334] "Generic (PLEG): container finished" podID="4b264ec9-951e-4928-b43c-5c045b7681f9" containerID="210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3" exitCode=0 Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.335383 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" event={"ID":"4b264ec9-951e-4928-b43c-5c045b7681f9","Type":"ContainerDied","Data":"210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3"} Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.342454 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.342518 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.342529 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.342565 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.342579 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:45Z","lastTransitionTime":"2025-09-30T13:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:45 crc kubenswrapper[4840]: E0930 13:56:45.354456 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5\\\",\\\"systemUUID\\\":\\\"f41e0351-4eda-411a-9a99-43a1952f3d34\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:45Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.354687 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:45Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.357968 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.358012 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.358023 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.358041 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.358055 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:45Z","lastTransitionTime":"2025-09-30T13:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.368899 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:45Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:45 crc kubenswrapper[4840]: E0930 13:56:45.371001 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5\\\",\\\"systemUUID\\\":\\\"f41e0351-4eda-411a-9a99-43a1952f3d34\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:45Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.376172 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.376213 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.376224 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.376242 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.376253 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:45Z","lastTransitionTime":"2025-09-30T13:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.383478 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f8cd5bd-4a7f-4a5b-ab00-0cc64de9de94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cba320f443ed90a4e755228b663a1da3096e7ea8e068671d1f6273544b6ef60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c507106f39ae2e8acd4268c4e8f28ffd88a0965e69544d8d8a8830e83eceaf9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d62fbefb67a24fedc5f13652596c76f5b288fec58d79b5209f49c79db139d79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\"\\\\nI0930 13:56:34.911386 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 13:56:34.920195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 13:56:34.920220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 13:56:34.920240 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 13:56:34.920248 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 13:56:34.927953 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 13:56:34.927977 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927984 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 13:56:34.927992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 13:56:34.927996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 13:56:34.928001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 13:56:34.928159 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0930 13:56:34.930945 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759240579\\\\\\\\\\\\\\\" (2025-09-30 13:56:19 +0000 UTC to 2025-10-30 13:56:20 +0000 UTC (now=2025-09-30 13:56:34.930900435 +0000 UTC))\\\\\\\"\\\\nF0930 13:56:34.930995 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1435e2e88d45e5561ff5863095d0f953ef1478d19d87691ab226cad7ae148f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:45Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:45 crc kubenswrapper[4840]: E0930 13:56:45.389874 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5\\\",\\\"systemUUID\\\":\\\"f41e0351-4eda-411a-9a99-43a1952f3d34\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:45Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:45 crc kubenswrapper[4840]: E0930 13:56:45.389980 4840 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.391962 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.392003 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.392013 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.392032 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.392041 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:45Z","lastTransitionTime":"2025-09-30T13:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.395940 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:45Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.409587 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:45Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.423273 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b16edc8f720d4296d7fe50fe407f5983ab975818ec6954d1f0e738e5596624a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:45Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.432231 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:45Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.449229 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:45Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.461029 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:45Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.474535 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:45Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.488001 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:45Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.495059 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.495086 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.495095 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.495109 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.495119 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:45Z","lastTransitionTime":"2025-09-30T13:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.499783 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:45Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.511300 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pj2zd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4de996b8-bda8-4142-b5e0-04ed9ae1f327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7286e2d685d6573993e2fa000cb9e08d85c6aa820af507327be52ca50fe58c7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pj2zd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:45Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.598487 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.598532 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.598544 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.598584 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.598597 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:45Z","lastTransitionTime":"2025-09-30T13:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.702355 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.702673 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.702682 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.702698 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.702708 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:45Z","lastTransitionTime":"2025-09-30T13:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.805668 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.805710 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.805718 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.805733 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.805743 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:45Z","lastTransitionTime":"2025-09-30T13:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.908767 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.908842 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.908866 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.908899 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:45 crc kubenswrapper[4840]: I0930 13:56:45.908917 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:45Z","lastTransitionTime":"2025-09-30T13:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.011946 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.011978 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.011989 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.012003 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.012012 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:46Z","lastTransitionTime":"2025-09-30T13:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.114456 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.114622 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.114695 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.114768 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.114838 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:46Z","lastTransitionTime":"2025-09-30T13:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.116178 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:56:46 crc kubenswrapper[4840]: E0930 13:56:46.116411 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.218053 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.218097 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.218106 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.218123 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.218133 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:46Z","lastTransitionTime":"2025-09-30T13:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.321060 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.321098 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.321106 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.321121 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.321132 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:46Z","lastTransitionTime":"2025-09-30T13:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.343808 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" event={"ID":"2ff48c28-d076-46e8-a93f-9630989f81e8","Type":"ContainerStarted","Data":"c8cb31464b02d3ac9f62610436fdaa00d009301848cf17a62ec1f7deeef08fbf"} Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.344062 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.344086 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.348110 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" event={"ID":"4b264ec9-951e-4928-b43c-5c045b7681f9","Type":"ContainerStarted","Data":"c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5"} Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.365492 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84052a424fd6279d70b54902ebdcf430c8aa53b02406b52fc261db9d4d88ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27a94b18137fa991459484406b1c5bd2b2dbb52f527d5f13156ba8e5b8d0146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc4e5d3625aa2d5f94307cedddf72cc48f17f327a044501144e35422d670d555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9549ce85242f1f19b3a52e603caec58bb4c3e2449eb50d8c4525f56dfc93e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a364a13c773fe00ee4e022a286821013d3a681acc16567e66b1c33c7e18ee3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cb31464b02d3ac9f62610436fdaa00d009301848cf17a62ec1f7deeef08fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:46Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.376252 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.380973 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:46Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.395818 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:46Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.410072 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:46Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.424053 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.424091 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.424101 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.424117 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.424127 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:46Z","lastTransitionTime":"2025-09-30T13:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.424868 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:46Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.440789 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pj2zd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4de996b8-bda8-4142-b5e0-04ed9ae1f327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7286e2d685d6573993e2fa000cb9e08d85c6aa820af507327be52ca50fe58c7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pj2zd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:46Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.454427 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:46Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.466309 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:46Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.479838 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f8cd5bd-4a7f-4a5b-ab00-0cc64de9de94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cba320f443ed90a4e755228b663a1da3096e7ea8e068671d1f6273544b6ef60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c507106f39ae2e8acd4268c4e8f28ffd88a0965e69544d8d8a8830e83eceaf9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d62fbefb67a24fedc5f13652596c76f5b288fec58d79b5209f49c79db139d79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\"\\\\nI0930 13:56:34.911386 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 13:56:34.920195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 13:56:34.920220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 13:56:34.920240 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 13:56:34.920248 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 13:56:34.927953 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 13:56:34.927977 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927984 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 13:56:34.927992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 13:56:34.927996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 13:56:34.928001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 13:56:34.928159 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0930 13:56:34.930945 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759240579\\\\\\\\\\\\\\\" (2025-09-30 13:56:19 +0000 UTC to 2025-10-30 13:56:20 +0000 UTC (now=2025-09-30 13:56:34.930900435 +0000 UTC))\\\\\\\"\\\\nF0930 13:56:34.930995 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1435e2e88d45e5561ff5863095d0f953ef1478d19d87691ab226cad7ae148f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:46Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.492768 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:46Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.506840 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:46Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.519697 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b16edc8f720d4296d7fe50fe407f5983ab975818ec6954d1f0e738e5596624a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:46Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.526365 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.526394 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.526406 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.526427 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.526440 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:46Z","lastTransitionTime":"2025-09-30T13:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.531988 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:46Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.545961 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:46Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.564118 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:46Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.576308 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:46Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.593676 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f8cd5bd-4a7f-4a5b-ab00-0cc64de9de94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cba320f443ed90a4e755228b663a1da3096e7ea8e068671d1f6273544b6ef60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c507106f39ae2e8acd4268c4e8f28ffd88a0965e69544d8d8a8830e83eceaf9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d62fbefb67a24fedc5f13652596c76f5b288fec58d79b5209f49c79db139d79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\"\\\\nI0930 13:56:34.911386 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 13:56:34.920195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 13:56:34.920220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 13:56:34.920240 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 13:56:34.920248 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 13:56:34.927953 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 13:56:34.927977 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927984 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 13:56:34.927992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 13:56:34.927996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 13:56:34.928001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 13:56:34.928159 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0930 13:56:34.930945 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759240579\\\\\\\\\\\\\\\" (2025-09-30 13:56:19 +0000 UTC to 2025-10-30 13:56:20 +0000 UTC (now=2025-09-30 13:56:34.930900435 +0000 UTC))\\\\\\\"\\\\nF0930 13:56:34.930995 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1435e2e88d45e5561ff5863095d0f953ef1478d19d87691ab226cad7ae148f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:46Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.606632 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:46Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.625250 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:46Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.628850 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.628909 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.628934 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.628961 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.628981 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:46Z","lastTransitionTime":"2025-09-30T13:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.638240 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b16edc8f720d4296d7fe50fe407f5983ab975818ec6954d1f0e738e5596624a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:46Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.660678 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84052a424fd6279d70b54902ebdcf430c8aa53b02406b52fc261db9d4d88ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27a94b18137fa991459484406b1c5bd2b2dbb52f527d5f13156ba8e5b8d0146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc4e5d3625aa2d5f94307cedddf72cc48f17f327a044501144e35422d670d555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9549ce85242f1f19b3a52e603caec58bb4c3e2449eb50d8c4525f56dfc93e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a364a13c773fe00ee4e022a286821013d3a681acc16567e66b1c33c7e18ee3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cb31464b02d3ac9f62610436fdaa00d009301848cf17a62ec1f7deeef08fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:46Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.675464 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:46Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.689773 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:46Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.701721 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:46Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.715950 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:46Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.729114 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pj2zd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4de996b8-bda8-4142-b5e0-04ed9ae1f327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7286e2d685d6573993e2fa000cb9e08d85c6aa820af507327be52ca50fe58c7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pj2zd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:46Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.731304 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.731347 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.731362 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.731384 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.731398 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:46Z","lastTransitionTime":"2025-09-30T13:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.761079 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.761200 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:56:46 crc kubenswrapper[4840]: E0930 13:56:46.761251 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:56:54.761226891 +0000 UTC m=+43.390313324 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.761297 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.761339 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.761378 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:56:46 crc kubenswrapper[4840]: E0930 13:56:46.761381 4840 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 13:56:46 crc kubenswrapper[4840]: E0930 13:56:46.761482 4840 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 13:56:46 crc kubenswrapper[4840]: E0930 13:56:46.761496 4840 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 13:56:46 crc kubenswrapper[4840]: E0930 13:56:46.761709 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-30 13:56:54.761696084 +0000 UTC m=+43.390782517 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 13:56:46 crc kubenswrapper[4840]: E0930 13:56:46.762048 4840 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 13:56:46 crc kubenswrapper[4840]: E0930 13:56:46.762069 4840 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 13:56:46 crc kubenswrapper[4840]: E0930 13:56:46.762080 4840 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 13:56:46 crc kubenswrapper[4840]: E0930 13:56:46.762111 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-30 13:56:54.762100995 +0000 UTC m=+43.391187428 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 13:56:46 crc kubenswrapper[4840]: E0930 13:56:46.761449 4840 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 13:56:46 crc kubenswrapper[4840]: E0930 13:56:46.762143 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 13:56:54.762135405 +0000 UTC m=+43.391221838 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 13:56:46 crc kubenswrapper[4840]: E0930 13:56:46.762312 4840 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 13:56:46 crc kubenswrapper[4840]: E0930 13:56:46.762371 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 13:56:54.762359121 +0000 UTC m=+43.391445554 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.833789 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.833831 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.833844 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.833863 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.833882 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:46Z","lastTransitionTime":"2025-09-30T13:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.936312 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.936349 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.936358 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.936374 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:46 crc kubenswrapper[4840]: I0930 13:56:46.936384 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:46Z","lastTransitionTime":"2025-09-30T13:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.038264 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.038303 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.038314 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.038331 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.038342 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:47Z","lastTransitionTime":"2025-09-30T13:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.116426 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.116426 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:56:47 crc kubenswrapper[4840]: E0930 13:56:47.116633 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:56:47 crc kubenswrapper[4840]: E0930 13:56:47.116729 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.140656 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.140703 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.140718 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.140739 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.140753 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:47Z","lastTransitionTime":"2025-09-30T13:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.243182 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.243230 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.243242 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.243262 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.243273 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:47Z","lastTransitionTime":"2025-09-30T13:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.346958 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.347013 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.347030 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.347055 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.347074 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:47Z","lastTransitionTime":"2025-09-30T13:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.355432 4840 generic.go:334] "Generic (PLEG): container finished" podID="4b264ec9-951e-4928-b43c-5c045b7681f9" containerID="c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5" exitCode=0 Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.355537 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" event={"ID":"4b264ec9-951e-4928-b43c-5c045b7681f9","Type":"ContainerDied","Data":"c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5"} Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.356174 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.369624 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:47Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.386466 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.390413 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f8cd5bd-4a7f-4a5b-ab00-0cc64de9de94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cba320f443ed90a4e755228b663a1da3096e7ea8e068671d1f6273544b6ef60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c507106f39ae2e8acd4268c4e8f28ffd88a0965e69544d8d8a8830e83eceaf9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d62fbefb67a24fedc5f13652596c76f5b288fec58d79b5209f49c79db139d79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\"\\\\nI0930 13:56:34.911386 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 13:56:34.920195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 13:56:34.920220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 13:56:34.920240 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 13:56:34.920248 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 13:56:34.927953 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 13:56:34.927977 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927984 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 13:56:34.927992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 13:56:34.927996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 13:56:34.928001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 13:56:34.928159 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0930 13:56:34.930945 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759240579\\\\\\\\\\\\\\\" (2025-09-30 13:56:19 +0000 UTC to 2025-10-30 13:56:20 +0000 UTC (now=2025-09-30 13:56:34.930900435 +0000 UTC))\\\\\\\"\\\\nF0930 13:56:34.930995 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1435e2e88d45e5561ff5863095d0f953ef1478d19d87691ab226cad7ae148f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:47Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.406709 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:47Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.425232 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:47Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.438028 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b16edc8f720d4296d7fe50fe407f5983ab975818ec6954d1f0e738e5596624a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:47Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.451584 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.451635 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.451645 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.451669 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.451684 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:47Z","lastTransitionTime":"2025-09-30T13:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.459436 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84052a424fd6279d70b54902ebdcf430c8aa53b02406b52fc261db9d4d88ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27a94b18137fa991459484406b1c5bd2b2dbb52f527d5f13156ba8e5b8d0146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc4e5d3625aa2d5f94307cedddf72cc48f17f327a044501144e35422d670d555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9549ce85242f1f19b3a52e603caec58bb4c3e2449eb50d8c4525f56dfc93e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a364a13c773fe00ee4e022a286821013d3a681acc16567e66b1c33c7e18ee3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cb31464b02d3ac9f62610436fdaa00d009301848cf17a62ec1f7deeef08fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:47Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.479062 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:47Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.496028 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:47Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.510713 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:47Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.523703 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:47Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.534902 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pj2zd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4de996b8-bda8-4142-b5e0-04ed9ae1f327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7286e2d685d6573993e2fa000cb9e08d85c6aa820af507327be52ca50fe58c7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pj2zd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:47Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.552026 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:47Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.554040 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.554075 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.554087 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.554106 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.554119 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:47Z","lastTransitionTime":"2025-09-30T13:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.567343 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:47Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.589468 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:47Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.603982 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:47Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.617971 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:47Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.629963 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pj2zd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4de996b8-bda8-4142-b5e0-04ed9ae1f327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7286e2d685d6573993e2fa000cb9e08d85c6aa820af507327be52ca50fe58c7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pj2zd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:47Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.644083 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:47Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.656618 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.656666 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.656676 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.656695 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.656707 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:47Z","lastTransitionTime":"2025-09-30T13:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.659509 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:47Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.675532 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f8cd5bd-4a7f-4a5b-ab00-0cc64de9de94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cba320f443ed90a4e755228b663a1da3096e7ea8e068671d1f6273544b6ef60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c507106f39ae2e8acd4268c4e8f28ffd88a0965e69544d8d8a8830e83eceaf9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d62fbefb67a24fedc5f13652596c76f5b288fec58d79b5209f49c79db139d79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\"\\\\nI0930 13:56:34.911386 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 13:56:34.920195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 13:56:34.920220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 13:56:34.920240 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 13:56:34.920248 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 13:56:34.927953 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 13:56:34.927977 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927984 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 13:56:34.927992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 13:56:34.927996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 13:56:34.928001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 13:56:34.928159 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0930 13:56:34.930945 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759240579\\\\\\\\\\\\\\\" (2025-09-30 13:56:19 +0000 UTC to 2025-10-30 13:56:20 +0000 UTC (now=2025-09-30 13:56:34.930900435 +0000 UTC))\\\\\\\"\\\\nF0930 13:56:34.930995 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1435e2e88d45e5561ff5863095d0f953ef1478d19d87691ab226cad7ae148f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:47Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.688286 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:47Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.701344 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:47Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.712655 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b16edc8f720d4296d7fe50fe407f5983ab975818ec6954d1f0e738e5596624a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:47Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.722198 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:47Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.738860 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84052a424fd6279d70b54902ebdcf430c8aa53b02406b52fc261db9d4d88ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27a94b18137fa991459484406b1c5bd2b2dbb52f527d5f13156ba8e5b8d0146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc4e5d3625aa2d5f94307cedddf72cc48f17f327a044501144e35422d670d555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9549ce85242f1f19b3a52e603caec58bb4c3e2449eb50d8c4525f56dfc93e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a364a13c773fe00ee4e022a286821013d3a681acc16567e66b1c33c7e18ee3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cb31464b02d3ac9f62610436fdaa00d009301848cf17a62ec1f7deeef08fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:47Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.753357 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:47Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.760070 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.760106 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.760117 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.760135 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.760145 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:47Z","lastTransitionTime":"2025-09-30T13:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.863200 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.863243 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.863255 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.863274 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.863286 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:47Z","lastTransitionTime":"2025-09-30T13:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.965104 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.965173 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.965192 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.965221 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:47 crc kubenswrapper[4840]: I0930 13:56:47.965242 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:47Z","lastTransitionTime":"2025-09-30T13:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.067643 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.067677 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.067686 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.067701 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.067711 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:48Z","lastTransitionTime":"2025-09-30T13:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.115773 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:56:48 crc kubenswrapper[4840]: E0930 13:56:48.115965 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.170502 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.170562 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.170572 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.170589 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.170600 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:48Z","lastTransitionTime":"2025-09-30T13:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.273713 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.273774 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.273792 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.273818 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.273836 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:48Z","lastTransitionTime":"2025-09-30T13:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.366692 4840 generic.go:334] "Generic (PLEG): container finished" podID="4b264ec9-951e-4928-b43c-5c045b7681f9" containerID="a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e" exitCode=0 Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.367637 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" event={"ID":"4b264ec9-951e-4928-b43c-5c045b7681f9","Type":"ContainerDied","Data":"a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e"} Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.376161 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.376196 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.376206 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.376227 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.376239 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:48Z","lastTransitionTime":"2025-09-30T13:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.385853 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f8cd5bd-4a7f-4a5b-ab00-0cc64de9de94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cba320f443ed90a4e755228b663a1da3096e7ea8e068671d1f6273544b6ef60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c507106f39ae2e8acd4268c4e8f28ffd88a0965e69544d8d8a8830e83eceaf9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d62fbefb67a24fedc5f13652596c76f5b288fec58d79b5209f49c79db139d79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\"\\\\nI0930 13:56:34.911386 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 13:56:34.920195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 13:56:34.920220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 13:56:34.920240 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 13:56:34.920248 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 13:56:34.927953 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 13:56:34.927977 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927984 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 13:56:34.927992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 13:56:34.927996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 13:56:34.928001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 13:56:34.928159 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0930 13:56:34.930945 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759240579\\\\\\\\\\\\\\\" (2025-09-30 13:56:19 +0000 UTC to 2025-10-30 13:56:20 +0000 UTC (now=2025-09-30 13:56:34.930900435 +0000 UTC))\\\\\\\"\\\\nF0930 13:56:34.930995 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1435e2e88d45e5561ff5863095d0f953ef1478d19d87691ab226cad7ae148f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:48Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.400207 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:48Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.415821 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:48Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.428603 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b16edc8f720d4296d7fe50fe407f5983ab975818ec6954d1f0e738e5596624a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:48Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.439511 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:48Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.455786 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84052a424fd6279d70b54902ebdcf430c8aa53b02406b52fc261db9d4d88ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27a94b18137fa991459484406b1c5bd2b2dbb52f527d5f13156ba8e5b8d0146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc4e5d3625aa2d5f94307cedddf72cc48f17f327a044501144e35422d670d555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9549ce85242f1f19b3a52e603caec58bb4c3e2449eb50d8c4525f56dfc93e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a364a13c773fe00ee4e022a286821013d3a681acc16567e66b1c33c7e18ee3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cb31464b02d3ac9f62610436fdaa00d009301848cf17a62ec1f7deeef08fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:48Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.469357 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:48Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.479936 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.479988 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.479998 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.480011 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.480021 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:48Z","lastTransitionTime":"2025-09-30T13:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.487977 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:48Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.504476 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:48Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.517395 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:48Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.527402 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pj2zd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4de996b8-bda8-4142-b5e0-04ed9ae1f327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7286e2d685d6573993e2fa000cb9e08d85c6aa820af507327be52ca50fe58c7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pj2zd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:48Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.538929 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:48Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.553954 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:48Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.583350 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.583698 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.583711 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.583740 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.583753 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:48Z","lastTransitionTime":"2025-09-30T13:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.686008 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.686072 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.686086 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.686106 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.686119 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:48Z","lastTransitionTime":"2025-09-30T13:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.788436 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.788487 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.788499 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.788520 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.788540 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:48Z","lastTransitionTime":"2025-09-30T13:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.891363 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.891421 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.891434 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.891455 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.891468 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:48Z","lastTransitionTime":"2025-09-30T13:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.994687 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.994741 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.994752 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.994773 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:48 crc kubenswrapper[4840]: I0930 13:56:48.994784 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:48Z","lastTransitionTime":"2025-09-30T13:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.061630 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5"] Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.062165 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" Sep 30 13:56:49 crc kubenswrapper[4840]: W0930 13:56:49.065192 4840 reflector.go:561] object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert": failed to list *v1.Secret: secrets "ovn-control-plane-metrics-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Sep 30 13:56:49 crc kubenswrapper[4840]: E0930 13:56:49.065247 4840 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"ovn-control-plane-metrics-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"ovn-control-plane-metrics-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Sep 30 13:56:49 crc kubenswrapper[4840]: W0930 13:56:49.066254 4840 reflector.go:561] object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd": failed to list *v1.Secret: secrets "ovn-kubernetes-control-plane-dockercfg-gs7dd" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Sep 30 13:56:49 crc kubenswrapper[4840]: E0930 13:56:49.066287 4840 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-control-plane-dockercfg-gs7dd\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"ovn-kubernetes-control-plane-dockercfg-gs7dd\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.092619 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.097796 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.097839 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.097853 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.097874 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.097886 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:49Z","lastTransitionTime":"2025-09-30T13:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.108602 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.115639 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.115668 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:56:49 crc kubenswrapper[4840]: E0930 13:56:49.115779 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:56:49 crc kubenswrapper[4840]: E0930 13:56:49.115883 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.122940 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.133031 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pj2zd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4de996b8-bda8-4142-b5e0-04ed9ae1f327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7286e2d685d6573993e2fa000cb9e08d85c6aa820af507327be52ca50fe58c7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pj2zd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.147897 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.162005 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.175098 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f8cd5bd-4a7f-4a5b-ab00-0cc64de9de94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cba320f443ed90a4e755228b663a1da3096e7ea8e068671d1f6273544b6ef60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c507106f39ae2e8acd4268c4e8f28ffd88a0965e69544d8d8a8830e83eceaf9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d62fbefb67a24fedc5f13652596c76f5b288fec58d79b5209f49c79db139d79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\"\\\\nI0930 13:56:34.911386 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 13:56:34.920195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 13:56:34.920220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 13:56:34.920240 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 13:56:34.920248 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 13:56:34.927953 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 13:56:34.927977 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927984 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 13:56:34.927992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 13:56:34.927996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 13:56:34.928001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 13:56:34.928159 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0930 13:56:34.930945 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759240579\\\\\\\\\\\\\\\" (2025-09-30 13:56:19 +0000 UTC to 2025-10-30 13:56:20 +0000 UTC (now=2025-09-30 13:56:34.930900435 +0000 UTC))\\\\\\\"\\\\nF0930 13:56:34.930995 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1435e2e88d45e5561ff5863095d0f953ef1478d19d87691ab226cad7ae148f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.189367 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.190904 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8c2d8050-9531-4413-9d18-14bf3be65acf-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-7nvs5\" (UID: \"8c2d8050-9531-4413-9d18-14bf3be65acf\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.190943 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wr4j9\" (UniqueName: \"kubernetes.io/projected/8c2d8050-9531-4413-9d18-14bf3be65acf-kube-api-access-wr4j9\") pod \"ovnkube-control-plane-749d76644c-7nvs5\" (UID: \"8c2d8050-9531-4413-9d18-14bf3be65acf\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.190994 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8c2d8050-9531-4413-9d18-14bf3be65acf-env-overrides\") pod \"ovnkube-control-plane-749d76644c-7nvs5\" (UID: \"8c2d8050-9531-4413-9d18-14bf3be65acf\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.191021 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8c2d8050-9531-4413-9d18-14bf3be65acf-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-7nvs5\" (UID: \"8c2d8050-9531-4413-9d18-14bf3be65acf\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.200512 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.200588 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.200600 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.200619 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.200632 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:49Z","lastTransitionTime":"2025-09-30T13:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.206444 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.223513 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b16edc8f720d4296d7fe50fe407f5983ab975818ec6954d1f0e738e5596624a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.234896 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.247888 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c2d8050-9531-4413-9d18-14bf3be65acf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7nvs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.268835 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84052a424fd6279d70b54902ebdcf430c8aa53b02406b52fc261db9d4d88ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27a94b18137fa991459484406b1c5bd2b2dbb52f527d5f13156ba8e5b8d0146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc4e5d3625aa2d5f94307cedddf72cc48f17f327a044501144e35422d670d555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9549ce85242f1f19b3a52e603caec58bb4c3e2449eb50d8c4525f56dfc93e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a364a13c773fe00ee4e022a286821013d3a681acc16567e66b1c33c7e18ee3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cb31464b02d3ac9f62610436fdaa00d009301848cf17a62ec1f7deeef08fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.285346 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.291660 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8c2d8050-9531-4413-9d18-14bf3be65acf-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-7nvs5\" (UID: \"8c2d8050-9531-4413-9d18-14bf3be65acf\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.291721 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8c2d8050-9531-4413-9d18-14bf3be65acf-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-7nvs5\" (UID: \"8c2d8050-9531-4413-9d18-14bf3be65acf\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.291738 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wr4j9\" (UniqueName: \"kubernetes.io/projected/8c2d8050-9531-4413-9d18-14bf3be65acf-kube-api-access-wr4j9\") pod \"ovnkube-control-plane-749d76644c-7nvs5\" (UID: \"8c2d8050-9531-4413-9d18-14bf3be65acf\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.291776 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8c2d8050-9531-4413-9d18-14bf3be65acf-env-overrides\") pod \"ovnkube-control-plane-749d76644c-7nvs5\" (UID: \"8c2d8050-9531-4413-9d18-14bf3be65acf\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.292239 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8c2d8050-9531-4413-9d18-14bf3be65acf-env-overrides\") pod \"ovnkube-control-plane-749d76644c-7nvs5\" (UID: \"8c2d8050-9531-4413-9d18-14bf3be65acf\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.292495 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8c2d8050-9531-4413-9d18-14bf3be65acf-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-7nvs5\" (UID: \"8c2d8050-9531-4413-9d18-14bf3be65acf\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.303066 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.303093 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.303102 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.303117 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.303125 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:49Z","lastTransitionTime":"2025-09-30T13:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.312565 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wr4j9\" (UniqueName: \"kubernetes.io/projected/8c2d8050-9531-4413-9d18-14bf3be65acf-kube-api-access-wr4j9\") pod \"ovnkube-control-plane-749d76644c-7nvs5\" (UID: \"8c2d8050-9531-4413-9d18-14bf3be65acf\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.374228 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" event={"ID":"4b264ec9-951e-4928-b43c-5c045b7681f9","Type":"ContainerStarted","Data":"c8d388c29baf0ed22d5c9240f2585e4ca133de265227af6c802a169da6650f95"} Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.396520 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84052a424fd6279d70b54902ebdcf430c8aa53b02406b52fc261db9d4d88ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27a94b18137fa991459484406b1c5bd2b2dbb52f527d5f13156ba8e5b8d0146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc4e5d3625aa2d5f94307cedddf72cc48f17f327a044501144e35422d670d555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9549ce85242f1f19b3a52e603caec58bb4c3e2449eb50d8c4525f56dfc93e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a364a13c773fe00ee4e022a286821013d3a681acc16567e66b1c33c7e18ee3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cb31464b02d3ac9f62610436fdaa00d009301848cf17a62ec1f7deeef08fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.405719 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.405752 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.405762 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.405779 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.405792 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:49Z","lastTransitionTime":"2025-09-30T13:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.410818 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.427009 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.440491 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.456635 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.468719 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pj2zd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4de996b8-bda8-4142-b5e0-04ed9ae1f327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7286e2d685d6573993e2fa000cb9e08d85c6aa820af507327be52ca50fe58c7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pj2zd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.481924 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.496417 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.508224 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.508289 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.508300 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.508317 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.508352 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:49Z","lastTransitionTime":"2025-09-30T13:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.511179 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f8cd5bd-4a7f-4a5b-ab00-0cc64de9de94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cba320f443ed90a4e755228b663a1da3096e7ea8e068671d1f6273544b6ef60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c507106f39ae2e8acd4268c4e8f28ffd88a0965e69544d8d8a8830e83eceaf9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d62fbefb67a24fedc5f13652596c76f5b288fec58d79b5209f49c79db139d79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\"\\\\nI0930 13:56:34.911386 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 13:56:34.920195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 13:56:34.920220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 13:56:34.920240 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 13:56:34.920248 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 13:56:34.927953 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 13:56:34.927977 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927984 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 13:56:34.927992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 13:56:34.927996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 13:56:34.928001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 13:56:34.928159 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0930 13:56:34.930945 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759240579\\\\\\\\\\\\\\\" (2025-09-30 13:56:19 +0000 UTC to 2025-10-30 13:56:20 +0000 UTC (now=2025-09-30 13:56:34.930900435 +0000 UTC))\\\\\\\"\\\\nF0930 13:56:34.930995 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1435e2e88d45e5561ff5863095d0f953ef1478d19d87691ab226cad7ae148f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.521648 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.534253 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8d388c29baf0ed22d5c9240f2585e4ca133de265227af6c802a169da6650f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.546274 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b16edc8f720d4296d7fe50fe407f5983ab975818ec6954d1f0e738e5596624a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.556650 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.567541 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c2d8050-9531-4413-9d18-14bf3be65acf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7nvs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.610175 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.610204 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.610213 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.610228 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.610238 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:49Z","lastTransitionTime":"2025-09-30T13:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.712355 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.712401 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.712416 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.712440 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.712456 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:49Z","lastTransitionTime":"2025-09-30T13:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.815378 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.815464 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.815495 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.815527 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.815547 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:49Z","lastTransitionTime":"2025-09-30T13:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.828925 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-4gp5x"] Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.829348 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:56:49 crc kubenswrapper[4840]: E0930 13:56:49.829412 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.848116 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84052a424fd6279d70b54902ebdcf430c8aa53b02406b52fc261db9d4d88ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27a94b18137fa991459484406b1c5bd2b2dbb52f527d5f13156ba8e5b8d0146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc4e5d3625aa2d5f94307cedddf72cc48f17f327a044501144e35422d670d555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9549ce85242f1f19b3a52e603caec58bb4c3e2449eb50d8c4525f56dfc93e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a364a13c773fe00ee4e022a286821013d3a681acc16567e66b1c33c7e18ee3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cb31464b02d3ac9f62610436fdaa00d009301848cf17a62ec1f7deeef08fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.860825 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.873252 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.883642 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pj2zd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4de996b8-bda8-4142-b5e0-04ed9ae1f327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7286e2d685d6573993e2fa000cb9e08d85c6aa820af507327be52ca50fe58c7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pj2zd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.892601 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.897983 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvf8m\" (UniqueName: \"kubernetes.io/projected/1491f559-bc12-4afd-a40c-4eaa40d920a8-kube-api-access-mvf8m\") pod \"network-metrics-daemon-4gp5x\" (UID: \"1491f559-bc12-4afd-a40c-4eaa40d920a8\") " pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.898056 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1491f559-bc12-4afd-a40c-4eaa40d920a8-metrics-certs\") pod \"network-metrics-daemon-4gp5x\" (UID: \"1491f559-bc12-4afd-a40c-4eaa40d920a8\") " pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.900540 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.915918 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.917943 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.917973 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.917986 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.918006 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.918017 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:49Z","lastTransitionTime":"2025-09-30T13:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.936724 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.953305 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.967186 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8d388c29baf0ed22d5c9240f2585e4ca133de265227af6c802a169da6650f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.977861 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b16edc8f720d4296d7fe50fe407f5983ab975818ec6954d1f0e738e5596624a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.986831 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.996479 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c2d8050-9531-4413-9d18-14bf3be65acf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7nvs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:49Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.998841 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvf8m\" (UniqueName: \"kubernetes.io/projected/1491f559-bc12-4afd-a40c-4eaa40d920a8-kube-api-access-mvf8m\") pod \"network-metrics-daemon-4gp5x\" (UID: \"1491f559-bc12-4afd-a40c-4eaa40d920a8\") " pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:56:49 crc kubenswrapper[4840]: I0930 13:56:49.998964 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1491f559-bc12-4afd-a40c-4eaa40d920a8-metrics-certs\") pod \"network-metrics-daemon-4gp5x\" (UID: \"1491f559-bc12-4afd-a40c-4eaa40d920a8\") " pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:56:49 crc kubenswrapper[4840]: E0930 13:56:49.999104 4840 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 13:56:49 crc kubenswrapper[4840]: E0930 13:56:49.999186 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1491f559-bc12-4afd-a40c-4eaa40d920a8-metrics-certs podName:1491f559-bc12-4afd-a40c-4eaa40d920a8 nodeName:}" failed. No retries permitted until 2025-09-30 13:56:50.499164972 +0000 UTC m=+39.128251415 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1491f559-bc12-4afd-a40c-4eaa40d920a8-metrics-certs") pod "network-metrics-daemon-4gp5x" (UID: "1491f559-bc12-4afd-a40c-4eaa40d920a8") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.006091 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gp5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1491f559-bc12-4afd-a40c-4eaa40d920a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gp5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:50Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.014993 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvf8m\" (UniqueName: \"kubernetes.io/projected/1491f559-bc12-4afd-a40c-4eaa40d920a8-kube-api-access-mvf8m\") pod \"network-metrics-daemon-4gp5x\" (UID: \"1491f559-bc12-4afd-a40c-4eaa40d920a8\") " pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.017225 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f8cd5bd-4a7f-4a5b-ab00-0cc64de9de94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cba320f443ed90a4e755228b663a1da3096e7ea8e068671d1f6273544b6ef60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c507106f39ae2e8acd4268c4e8f28ffd88a0965e69544d8d8a8830e83eceaf9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d62fbefb67a24fedc5f13652596c76f5b288fec58d79b5209f49c79db139d79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\"\\\\nI0930 13:56:34.911386 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 13:56:34.920195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 13:56:34.920220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 13:56:34.920240 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 13:56:34.920248 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 13:56:34.927953 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 13:56:34.927977 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927984 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 13:56:34.927992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 13:56:34.927996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 13:56:34.928001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 13:56:34.928159 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0930 13:56:34.930945 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759240579\\\\\\\\\\\\\\\" (2025-09-30 13:56:19 +0000 UTC to 2025-10-30 13:56:20 +0000 UTC (now=2025-09-30 13:56:34.930900435 +0000 UTC))\\\\\\\"\\\\nF0930 13:56:34.930995 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1435e2e88d45e5561ff5863095d0f953ef1478d19d87691ab226cad7ae148f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:50Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.019782 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.020026 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.020043 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.020068 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.020114 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:50Z","lastTransitionTime":"2025-09-30T13:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.028833 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:50Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.115465 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:56:50 crc kubenswrapper[4840]: E0930 13:56:50.115676 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.122866 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.122936 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.122953 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.122980 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.122997 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:50Z","lastTransitionTime":"2025-09-30T13:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.158180 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.165864 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8c2d8050-9531-4413-9d18-14bf3be65acf-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-7nvs5\" (UID: \"8c2d8050-9531-4413-9d18-14bf3be65acf\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.226388 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.226439 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.226454 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.226482 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.226494 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:50Z","lastTransitionTime":"2025-09-30T13:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.274793 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" Sep 30 13:56:50 crc kubenswrapper[4840]: W0930 13:56:50.293646 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c2d8050_9531_4413_9d18_14bf3be65acf.slice/crio-542af88040374e26dd03b6d3f65a6b673fa09587fc0c99a57a74689d46a5f529 WatchSource:0}: Error finding container 542af88040374e26dd03b6d3f65a6b673fa09587fc0c99a57a74689d46a5f529: Status 404 returned error can't find the container with id 542af88040374e26dd03b6d3f65a6b673fa09587fc0c99a57a74689d46a5f529 Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.329128 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.329169 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.329181 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.329199 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.329211 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:50Z","lastTransitionTime":"2025-09-30T13:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.378536 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" event={"ID":"8c2d8050-9531-4413-9d18-14bf3be65acf","Type":"ContainerStarted","Data":"542af88040374e26dd03b6d3f65a6b673fa09587fc0c99a57a74689d46a5f529"} Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.380436 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kfx69_2ff48c28-d076-46e8-a93f-9630989f81e8/ovnkube-controller/0.log" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.383213 4840 generic.go:334] "Generic (PLEG): container finished" podID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerID="c8cb31464b02d3ac9f62610436fdaa00d009301848cf17a62ec1f7deeef08fbf" exitCode=1 Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.383254 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" event={"ID":"2ff48c28-d076-46e8-a93f-9630989f81e8","Type":"ContainerDied","Data":"c8cb31464b02d3ac9f62610436fdaa00d009301848cf17a62ec1f7deeef08fbf"} Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.384110 4840 scope.go:117] "RemoveContainer" containerID="c8cb31464b02d3ac9f62610436fdaa00d009301848cf17a62ec1f7deeef08fbf" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.402585 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84052a424fd6279d70b54902ebdcf430c8aa53b02406b52fc261db9d4d88ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27a94b18137fa991459484406b1c5bd2b2dbb52f527d5f13156ba8e5b8d0146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc4e5d3625aa2d5f94307cedddf72cc48f17f327a044501144e35422d670d555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9549ce85242f1f19b3a52e603caec58bb4c3e2449eb50d8c4525f56dfc93e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a364a13c773fe00ee4e022a286821013d3a681acc16567e66b1c33c7e18ee3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cb31464b02d3ac9f62610436fdaa00d009301848cf17a62ec1f7deeef08fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8cb31464b02d3ac9f62610436fdaa00d009301848cf17a62ec1f7deeef08fbf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"tor.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 13:56:49.801872 6077 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 13:56:49.802099 6077 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0930 13:56:49.802145 6077 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0930 13:56:49.802229 6077 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0930 13:56:49.802238 6077 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0930 13:56:49.802249 6077 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 13:56:49.802255 6077 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0930 13:56:49.802283 6077 factory.go:656] Stopping watch factory\\\\nI0930 13:56:49.802277 6077 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0930 13:56:49.802297 6077 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0930 13:56:49.802309 6077 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 13:56:49.802319 6077 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0930 13:56:49.802326 6077 handler.go:208] Removed *v1.Node event handler 2\\\\nI0930 13:56:49.802333 6077 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:50Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.416228 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:50Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.431602 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:50Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.432321 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.432351 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.432361 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.432378 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.432391 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:50Z","lastTransitionTime":"2025-09-30T13:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.443790 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:50Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.460449 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:50Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.472521 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pj2zd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4de996b8-bda8-4142-b5e0-04ed9ae1f327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7286e2d685d6573993e2fa000cb9e08d85c6aa820af507327be52ca50fe58c7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pj2zd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:50Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.485700 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:50Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.501869 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:50Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.504910 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1491f559-bc12-4afd-a40c-4eaa40d920a8-metrics-certs\") pod \"network-metrics-daemon-4gp5x\" (UID: \"1491f559-bc12-4afd-a40c-4eaa40d920a8\") " pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:56:50 crc kubenswrapper[4840]: E0930 13:56:50.505017 4840 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 13:56:50 crc kubenswrapper[4840]: E0930 13:56:50.505069 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1491f559-bc12-4afd-a40c-4eaa40d920a8-metrics-certs podName:1491f559-bc12-4afd-a40c-4eaa40d920a8 nodeName:}" failed. No retries permitted until 2025-09-30 13:56:51.505055745 +0000 UTC m=+40.134142168 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1491f559-bc12-4afd-a40c-4eaa40d920a8-metrics-certs") pod "network-metrics-daemon-4gp5x" (UID: "1491f559-bc12-4afd-a40c-4eaa40d920a8") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.512884 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c2d8050-9531-4413-9d18-14bf3be65acf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7nvs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:50Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.521525 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gp5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1491f559-bc12-4afd-a40c-4eaa40d920a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gp5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:50Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.533872 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f8cd5bd-4a7f-4a5b-ab00-0cc64de9de94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cba320f443ed90a4e755228b663a1da3096e7ea8e068671d1f6273544b6ef60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c507106f39ae2e8acd4268c4e8f28ffd88a0965e69544d8d8a8830e83eceaf9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d62fbefb67a24fedc5f13652596c76f5b288fec58d79b5209f49c79db139d79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\"\\\\nI0930 13:56:34.911386 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 13:56:34.920195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 13:56:34.920220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 13:56:34.920240 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 13:56:34.920248 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 13:56:34.927953 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 13:56:34.927977 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927984 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 13:56:34.927992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 13:56:34.927996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 13:56:34.928001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 13:56:34.928159 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0930 13:56:34.930945 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759240579\\\\\\\\\\\\\\\" (2025-09-30 13:56:19 +0000 UTC to 2025-10-30 13:56:20 +0000 UTC (now=2025-09-30 13:56:34.930900435 +0000 UTC))\\\\\\\"\\\\nF0930 13:56:34.930995 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1435e2e88d45e5561ff5863095d0f953ef1478d19d87691ab226cad7ae148f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:50Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.534717 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.534950 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.534963 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.534983 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.534994 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:50Z","lastTransitionTime":"2025-09-30T13:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.546701 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:50Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.563501 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8d388c29baf0ed22d5c9240f2585e4ca133de265227af6c802a169da6650f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:50Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.579401 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b16edc8f720d4296d7fe50fe407f5983ab975818ec6954d1f0e738e5596624a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:50Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.592790 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:50Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.637246 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.637302 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.637316 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.637336 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.637349 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:50Z","lastTransitionTime":"2025-09-30T13:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.739404 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.739452 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.739462 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.739481 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.739496 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:50Z","lastTransitionTime":"2025-09-30T13:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.842497 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.842597 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.842612 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.842635 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.842650 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:50Z","lastTransitionTime":"2025-09-30T13:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.944474 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.944522 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.944535 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.944580 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:50 crc kubenswrapper[4840]: I0930 13:56:50.944594 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:50Z","lastTransitionTime":"2025-09-30T13:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.047130 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.047181 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.047199 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.047222 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.047237 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:51Z","lastTransitionTime":"2025-09-30T13:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.116014 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:56:51 crc kubenswrapper[4840]: E0930 13:56:51.116627 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.116862 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.116979 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:56:51 crc kubenswrapper[4840]: E0930 13:56:51.117053 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:56:51 crc kubenswrapper[4840]: E0930 13:56:51.117259 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.150397 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.150764 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.150912 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.151057 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.151190 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:51Z","lastTransitionTime":"2025-09-30T13:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.253389 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.253426 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.253434 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.253450 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.253462 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:51Z","lastTransitionTime":"2025-09-30T13:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.356467 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.356520 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.356531 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.356572 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.356583 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:51Z","lastTransitionTime":"2025-09-30T13:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.459039 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.459069 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.459077 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.459092 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.459101 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:51Z","lastTransitionTime":"2025-09-30T13:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.516312 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1491f559-bc12-4afd-a40c-4eaa40d920a8-metrics-certs\") pod \"network-metrics-daemon-4gp5x\" (UID: \"1491f559-bc12-4afd-a40c-4eaa40d920a8\") " pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:56:51 crc kubenswrapper[4840]: E0930 13:56:51.516509 4840 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 13:56:51 crc kubenswrapper[4840]: E0930 13:56:51.516671 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1491f559-bc12-4afd-a40c-4eaa40d920a8-metrics-certs podName:1491f559-bc12-4afd-a40c-4eaa40d920a8 nodeName:}" failed. No retries permitted until 2025-09-30 13:56:53.516643376 +0000 UTC m=+42.145729879 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1491f559-bc12-4afd-a40c-4eaa40d920a8-metrics-certs") pod "network-metrics-daemon-4gp5x" (UID: "1491f559-bc12-4afd-a40c-4eaa40d920a8") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.561970 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.562046 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.562056 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.562073 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.562083 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:51Z","lastTransitionTime":"2025-09-30T13:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.665020 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.665073 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.665086 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.665106 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.665116 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:51Z","lastTransitionTime":"2025-09-30T13:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.768287 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.768376 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.768398 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.768429 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.768449 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:51Z","lastTransitionTime":"2025-09-30T13:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.870908 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.870955 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.870964 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.870983 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.870994 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:51Z","lastTransitionTime":"2025-09-30T13:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.974430 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.974505 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.974523 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.974597 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:51 crc kubenswrapper[4840]: I0930 13:56:51.974635 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:51Z","lastTransitionTime":"2025-09-30T13:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.077712 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.077749 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.077759 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.077774 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.077785 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:52Z","lastTransitionTime":"2025-09-30T13:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.116657 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:56:52 crc kubenswrapper[4840]: E0930 13:56:52.116796 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.135389 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.150701 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.166360 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.179468 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.179525 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.179537 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.179574 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.179587 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:52Z","lastTransitionTime":"2025-09-30T13:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.182761 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pj2zd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4de996b8-bda8-4142-b5e0-04ed9ae1f327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7286e2d685d6573993e2fa000cb9e08d85c6aa820af507327be52ca50fe58c7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pj2zd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.200191 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.223660 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.246655 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gp5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1491f559-bc12-4afd-a40c-4eaa40d920a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gp5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.273332 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f8cd5bd-4a7f-4a5b-ab00-0cc64de9de94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cba320f443ed90a4e755228b663a1da3096e7ea8e068671d1f6273544b6ef60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c507106f39ae2e8acd4268c4e8f28ffd88a0965e69544d8d8a8830e83eceaf9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d62fbefb67a24fedc5f13652596c76f5b288fec58d79b5209f49c79db139d79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\"\\\\nI0930 13:56:34.911386 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 13:56:34.920195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 13:56:34.920220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 13:56:34.920240 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 13:56:34.920248 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 13:56:34.927953 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 13:56:34.927977 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927984 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 13:56:34.927992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 13:56:34.927996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 13:56:34.928001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 13:56:34.928159 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0930 13:56:34.930945 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759240579\\\\\\\\\\\\\\\" (2025-09-30 13:56:19 +0000 UTC to 2025-10-30 13:56:20 +0000 UTC (now=2025-09-30 13:56:34.930900435 +0000 UTC))\\\\\\\"\\\\nF0930 13:56:34.930995 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1435e2e88d45e5561ff5863095d0f953ef1478d19d87691ab226cad7ae148f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.281287 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.281329 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.281342 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.281359 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.281370 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:52Z","lastTransitionTime":"2025-09-30T13:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.294240 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.322058 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8d388c29baf0ed22d5c9240f2585e4ca133de265227af6c802a169da6650f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.333633 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b16edc8f720d4296d7fe50fe407f5983ab975818ec6954d1f0e738e5596624a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.345047 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.354527 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c2d8050-9531-4413-9d18-14bf3be65acf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7nvs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.372640 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84052a424fd6279d70b54902ebdcf430c8aa53b02406b52fc261db9d4d88ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27a94b18137fa991459484406b1c5bd2b2dbb52f527d5f13156ba8e5b8d0146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc4e5d3625aa2d5f94307cedddf72cc48f17f327a044501144e35422d670d555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9549ce85242f1f19b3a52e603caec58bb4c3e2449eb50d8c4525f56dfc93e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a364a13c773fe00ee4e022a286821013d3a681acc16567e66b1c33c7e18ee3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cb31464b02d3ac9f62610436fdaa00d009301848cf17a62ec1f7deeef08fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8cb31464b02d3ac9f62610436fdaa00d009301848cf17a62ec1f7deeef08fbf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"tor.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 13:56:49.801872 6077 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 13:56:49.802099 6077 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0930 13:56:49.802145 6077 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0930 13:56:49.802229 6077 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0930 13:56:49.802238 6077 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0930 13:56:49.802249 6077 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 13:56:49.802255 6077 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0930 13:56:49.802283 6077 factory.go:656] Stopping watch factory\\\\nI0930 13:56:49.802277 6077 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0930 13:56:49.802297 6077 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0930 13:56:49.802309 6077 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 13:56:49.802319 6077 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0930 13:56:49.802326 6077 handler.go:208] Removed *v1.Node event handler 2\\\\nI0930 13:56:49.802333 6077 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.384223 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.384269 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.384280 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.384300 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.384315 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:52Z","lastTransitionTime":"2025-09-30T13:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.387009 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.390825 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" event={"ID":"8c2d8050-9531-4413-9d18-14bf3be65acf","Type":"ContainerStarted","Data":"6b5244dfa04a1f1804ef64dc65030d5eaedf679849eec6a9f0731e6cf147f83b"} Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.390875 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" event={"ID":"8c2d8050-9531-4413-9d18-14bf3be65acf","Type":"ContainerStarted","Data":"34a542c60726f160b423c3e546ff4769a4a7e4a2441b5a09099e4d8c688286c2"} Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.392512 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kfx69_2ff48c28-d076-46e8-a93f-9630989f81e8/ovnkube-controller/0.log" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.395121 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" event={"ID":"2ff48c28-d076-46e8-a93f-9630989f81e8","Type":"ContainerStarted","Data":"2a2322a3823793c1dd16af892010cd522aeb59d3ccb359e34c603a3ae0181d32"} Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.395701 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.406527 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.416630 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.427817 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gp5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1491f559-bc12-4afd-a40c-4eaa40d920a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gp5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.443128 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f8cd5bd-4a7f-4a5b-ab00-0cc64de9de94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cba320f443ed90a4e755228b663a1da3096e7ea8e068671d1f6273544b6ef60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c507106f39ae2e8acd4268c4e8f28ffd88a0965e69544d8d8a8830e83eceaf9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d62fbefb67a24fedc5f13652596c76f5b288fec58d79b5209f49c79db139d79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\"\\\\nI0930 13:56:34.911386 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 13:56:34.920195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 13:56:34.920220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 13:56:34.920240 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 13:56:34.920248 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 13:56:34.927953 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 13:56:34.927977 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927984 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 13:56:34.927992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 13:56:34.927996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 13:56:34.928001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 13:56:34.928159 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0930 13:56:34.930945 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759240579\\\\\\\\\\\\\\\" (2025-09-30 13:56:19 +0000 UTC to 2025-10-30 13:56:20 +0000 UTC (now=2025-09-30 13:56:34.930900435 +0000 UTC))\\\\\\\"\\\\nF0930 13:56:34.930995 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1435e2e88d45e5561ff5863095d0f953ef1478d19d87691ab226cad7ae148f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.461609 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.477255 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8d388c29baf0ed22d5c9240f2585e4ca133de265227af6c802a169da6650f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.487401 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.487446 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.487458 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.487473 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.487484 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:52Z","lastTransitionTime":"2025-09-30T13:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.492328 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b16edc8f720d4296d7fe50fe407f5983ab975818ec6954d1f0e738e5596624a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.506787 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.521781 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c2d8050-9531-4413-9d18-14bf3be65acf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34a542c60726f160b423c3e546ff4769a4a7e4a2441b5a09099e4d8c688286c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b5244dfa04a1f1804ef64dc65030d5eaedf679849eec6a9f0731e6cf147f83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7nvs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.540322 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84052a424fd6279d70b54902ebdcf430c8aa53b02406b52fc261db9d4d88ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27a94b18137fa991459484406b1c5bd2b2dbb52f527d5f13156ba8e5b8d0146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc4e5d3625aa2d5f94307cedddf72cc48f17f327a044501144e35422d670d555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9549ce85242f1f19b3a52e603caec58bb4c3e2449eb50d8c4525f56dfc93e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a364a13c773fe00ee4e022a286821013d3a681acc16567e66b1c33c7e18ee3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cb31464b02d3ac9f62610436fdaa00d009301848cf17a62ec1f7deeef08fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8cb31464b02d3ac9f62610436fdaa00d009301848cf17a62ec1f7deeef08fbf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"tor.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 13:56:49.801872 6077 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 13:56:49.802099 6077 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0930 13:56:49.802145 6077 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0930 13:56:49.802229 6077 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0930 13:56:49.802238 6077 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0930 13:56:49.802249 6077 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 13:56:49.802255 6077 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0930 13:56:49.802283 6077 factory.go:656] Stopping watch factory\\\\nI0930 13:56:49.802277 6077 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0930 13:56:49.802297 6077 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0930 13:56:49.802309 6077 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 13:56:49.802319 6077 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0930 13:56:49.802326 6077 handler.go:208] Removed *v1.Node event handler 2\\\\nI0930 13:56:49.802333 6077 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.552516 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.568232 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.581253 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.589317 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.589353 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.589366 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.589383 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.589395 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:52Z","lastTransitionTime":"2025-09-30T13:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.594441 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.604814 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pj2zd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4de996b8-bda8-4142-b5e0-04ed9ae1f327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7286e2d685d6573993e2fa000cb9e08d85c6aa820af507327be52ca50fe58c7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pj2zd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.617746 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.631454 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.642877 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.655823 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pj2zd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4de996b8-bda8-4142-b5e0-04ed9ae1f327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7286e2d685d6573993e2fa000cb9e08d85c6aa820af507327be52ca50fe58c7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pj2zd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.669926 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.684588 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.692259 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.692331 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.692344 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.692363 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.692376 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:52Z","lastTransitionTime":"2025-09-30T13:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.699700 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.715087 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c2d8050-9531-4413-9d18-14bf3be65acf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34a542c60726f160b423c3e546ff4769a4a7e4a2441b5a09099e4d8c688286c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b5244dfa04a1f1804ef64dc65030d5eaedf679849eec6a9f0731e6cf147f83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7nvs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.726762 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gp5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1491f559-bc12-4afd-a40c-4eaa40d920a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gp5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.743822 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f8cd5bd-4a7f-4a5b-ab00-0cc64de9de94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cba320f443ed90a4e755228b663a1da3096e7ea8e068671d1f6273544b6ef60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c507106f39ae2e8acd4268c4e8f28ffd88a0965e69544d8d8a8830e83eceaf9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d62fbefb67a24fedc5f13652596c76f5b288fec58d79b5209f49c79db139d79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\"\\\\nI0930 13:56:34.911386 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 13:56:34.920195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 13:56:34.920220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 13:56:34.920240 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 13:56:34.920248 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 13:56:34.927953 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 13:56:34.927977 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927984 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 13:56:34.927992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 13:56:34.927996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 13:56:34.928001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 13:56:34.928159 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0930 13:56:34.930945 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759240579\\\\\\\\\\\\\\\" (2025-09-30 13:56:19 +0000 UTC to 2025-10-30 13:56:20 +0000 UTC (now=2025-09-30 13:56:34.930900435 +0000 UTC))\\\\\\\"\\\\nF0930 13:56:34.930995 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1435e2e88d45e5561ff5863095d0f953ef1478d19d87691ab226cad7ae148f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.756893 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.770722 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8d388c29baf0ed22d5c9240f2585e4ca133de265227af6c802a169da6650f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.782116 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b16edc8f720d4296d7fe50fe407f5983ab975818ec6954d1f0e738e5596624a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.795101 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.795175 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.795189 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.795239 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.795250 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:52Z","lastTransitionTime":"2025-09-30T13:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.803623 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84052a424fd6279d70b54902ebdcf430c8aa53b02406b52fc261db9d4d88ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27a94b18137fa991459484406b1c5bd2b2dbb52f527d5f13156ba8e5b8d0146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc4e5d3625aa2d5f94307cedddf72cc48f17f327a044501144e35422d670d555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9549ce85242f1f19b3a52e603caec58bb4c3e2449eb50d8c4525f56dfc93e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a364a13c773fe00ee4e022a286821013d3a681acc16567e66b1c33c7e18ee3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2322a3823793c1dd16af892010cd522aeb59d3ccb359e34c603a3ae0181d32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8cb31464b02d3ac9f62610436fdaa00d009301848cf17a62ec1f7deeef08fbf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"tor.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 13:56:49.801872 6077 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 13:56:49.802099 6077 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0930 13:56:49.802145 6077 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0930 13:56:49.802229 6077 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0930 13:56:49.802238 6077 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0930 13:56:49.802249 6077 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 13:56:49.802255 6077 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0930 13:56:49.802283 6077 factory.go:656] Stopping watch factory\\\\nI0930 13:56:49.802277 6077 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0930 13:56:49.802297 6077 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0930 13:56:49.802309 6077 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 13:56:49.802319 6077 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0930 13:56:49.802326 6077 handler.go:208] Removed *v1.Node event handler 2\\\\nI0930 13:56:49.802333 6077 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.824447 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.897650 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.897690 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.897701 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.897742 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:52 crc kubenswrapper[4840]: I0930 13:56:52.897754 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:52Z","lastTransitionTime":"2025-09-30T13:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.000343 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.000391 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.000399 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.000414 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.000424 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:53Z","lastTransitionTime":"2025-09-30T13:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.103429 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.103490 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.103510 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.103534 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.103595 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:53Z","lastTransitionTime":"2025-09-30T13:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.115879 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.115875 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:56:53 crc kubenswrapper[4840]: E0930 13:56:53.116033 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.116119 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:56:53 crc kubenswrapper[4840]: E0930 13:56:53.116700 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:56:53 crc kubenswrapper[4840]: E0930 13:56:53.116854 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.117057 4840 scope.go:117] "RemoveContainer" containerID="fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.205782 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.205813 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.205821 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.205836 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.205845 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:53Z","lastTransitionTime":"2025-09-30T13:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.308428 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.308467 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.308476 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.308490 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.308500 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:53Z","lastTransitionTime":"2025-09-30T13:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.399634 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.401451 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"61317d26f40a4361a289779bb97b235525b8dddae383f63abfdcab634b633479"} Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.401851 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.403020 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kfx69_2ff48c28-d076-46e8-a93f-9630989f81e8/ovnkube-controller/1.log" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.403453 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kfx69_2ff48c28-d076-46e8-a93f-9630989f81e8/ovnkube-controller/0.log" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.406178 4840 generic.go:334] "Generic (PLEG): container finished" podID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerID="2a2322a3823793c1dd16af892010cd522aeb59d3ccb359e34c603a3ae0181d32" exitCode=1 Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.406249 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" event={"ID":"2ff48c28-d076-46e8-a93f-9630989f81e8","Type":"ContainerDied","Data":"2a2322a3823793c1dd16af892010cd522aeb59d3ccb359e34c603a3ae0181d32"} Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.406282 4840 scope.go:117] "RemoveContainer" containerID="c8cb31464b02d3ac9f62610436fdaa00d009301848cf17a62ec1f7deeef08fbf" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.406826 4840 scope.go:117] "RemoveContainer" containerID="2a2322a3823793c1dd16af892010cd522aeb59d3ccb359e34c603a3ae0181d32" Sep 30 13:56:53 crc kubenswrapper[4840]: E0930 13:56:53.406957 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-kfx69_openshift-ovn-kubernetes(2ff48c28-d076-46e8-a93f-9630989f81e8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.410589 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.410625 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.410634 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.410646 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.410656 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:53Z","lastTransitionTime":"2025-09-30T13:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.425673 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84052a424fd6279d70b54902ebdcf430c8aa53b02406b52fc261db9d4d88ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27a94b18137fa991459484406b1c5bd2b2dbb52f527d5f13156ba8e5b8d0146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc4e5d3625aa2d5f94307cedddf72cc48f17f327a044501144e35422d670d555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9549ce85242f1f19b3a52e603caec58bb4c3e2449eb50d8c4525f56dfc93e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a364a13c773fe00ee4e022a286821013d3a681acc16567e66b1c33c7e18ee3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2322a3823793c1dd16af892010cd522aeb59d3ccb359e34c603a3ae0181d32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8cb31464b02d3ac9f62610436fdaa00d009301848cf17a62ec1f7deeef08fbf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"tor.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 13:56:49.801872 6077 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 13:56:49.802099 6077 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0930 13:56:49.802145 6077 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0930 13:56:49.802229 6077 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0930 13:56:49.802238 6077 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0930 13:56:49.802249 6077 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 13:56:49.802255 6077 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0930 13:56:49.802283 6077 factory.go:656] Stopping watch factory\\\\nI0930 13:56:49.802277 6077 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0930 13:56:49.802297 6077 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0930 13:56:49.802309 6077 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 13:56:49.802319 6077 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0930 13:56:49.802326 6077 handler.go:208] Removed *v1.Node event handler 2\\\\nI0930 13:56:49.802333 6077 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:53Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.440227 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:53Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.459078 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:53Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.470810 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pj2zd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4de996b8-bda8-4142-b5e0-04ed9ae1f327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7286e2d685d6573993e2fa000cb9e08d85c6aa820af507327be52ca50fe58c7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pj2zd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:53Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.485825 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:53Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.502366 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:53Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.513371 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.513425 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.513438 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.513458 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.513484 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:53Z","lastTransitionTime":"2025-09-30T13:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.515590 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:53Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.529235 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:53Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.536054 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1491f559-bc12-4afd-a40c-4eaa40d920a8-metrics-certs\") pod \"network-metrics-daemon-4gp5x\" (UID: \"1491f559-bc12-4afd-a40c-4eaa40d920a8\") " pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:56:53 crc kubenswrapper[4840]: E0930 13:56:53.536154 4840 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 13:56:53 crc kubenswrapper[4840]: E0930 13:56:53.536194 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1491f559-bc12-4afd-a40c-4eaa40d920a8-metrics-certs podName:1491f559-bc12-4afd-a40c-4eaa40d920a8 nodeName:}" failed. No retries permitted until 2025-09-30 13:56:57.536181881 +0000 UTC m=+46.165268304 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1491f559-bc12-4afd-a40c-4eaa40d920a8-metrics-certs") pod "network-metrics-daemon-4gp5x" (UID: "1491f559-bc12-4afd-a40c-4eaa40d920a8") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.551840 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8d388c29baf0ed22d5c9240f2585e4ca133de265227af6c802a169da6650f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:53Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.566198 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b16edc8f720d4296d7fe50fe407f5983ab975818ec6954d1f0e738e5596624a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:53Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.580589 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:53Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.593077 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c2d8050-9531-4413-9d18-14bf3be65acf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34a542c60726f160b423c3e546ff4769a4a7e4a2441b5a09099e4d8c688286c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b5244dfa04a1f1804ef64dc65030d5eaedf679849eec6a9f0731e6cf147f83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7nvs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:53Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.606273 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gp5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1491f559-bc12-4afd-a40c-4eaa40d920a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gp5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:53Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.616166 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.616205 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.616217 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.616231 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.616243 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:53Z","lastTransitionTime":"2025-09-30T13:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.620606 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f8cd5bd-4a7f-4a5b-ab00-0cc64de9de94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cba320f443ed90a4e755228b663a1da3096e7ea8e068671d1f6273544b6ef60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c507106f39ae2e8acd4268c4e8f28ffd88a0965e69544d8d8a8830e83eceaf9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d62fbefb67a24fedc5f13652596c76f5b288fec58d79b5209f49c79db139d79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61317d26f40a4361a289779bb97b235525b8dddae383f63abfdcab634b633479\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\"\\\\nI0930 13:56:34.911386 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 13:56:34.920195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 13:56:34.920220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 13:56:34.920240 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 13:56:34.920248 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 13:56:34.927953 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 13:56:34.927977 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927984 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 13:56:34.927992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 13:56:34.927996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 13:56:34.928001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 13:56:34.928159 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0930 13:56:34.930945 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759240579\\\\\\\\\\\\\\\" (2025-09-30 13:56:19 +0000 UTC to 2025-10-30 13:56:20 +0000 UTC (now=2025-09-30 13:56:34.930900435 +0000 UTC))\\\\\\\"\\\\nF0930 13:56:34.930995 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1435e2e88d45e5561ff5863095d0f953ef1478d19d87691ab226cad7ae148f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:53Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.638085 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:53Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.653806 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:53Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.666803 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:53Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.677043 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pj2zd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4de996b8-bda8-4142-b5e0-04ed9ae1f327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7286e2d685d6573993e2fa000cb9e08d85c6aa820af507327be52ca50fe58c7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pj2zd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:53Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.691613 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:53Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.703444 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:53Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.718001 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.718050 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.718062 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.718079 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.718091 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:53Z","lastTransitionTime":"2025-09-30T13:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.718593 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:53Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.732119 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:53Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.747117 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8d388c29baf0ed22d5c9240f2585e4ca133de265227af6c802a169da6650f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:53Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.759365 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b16edc8f720d4296d7fe50fe407f5983ab975818ec6954d1f0e738e5596624a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:53Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.769589 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:53Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.780408 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c2d8050-9531-4413-9d18-14bf3be65acf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34a542c60726f160b423c3e546ff4769a4a7e4a2441b5a09099e4d8c688286c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b5244dfa04a1f1804ef64dc65030d5eaedf679849eec6a9f0731e6cf147f83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7nvs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:53Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.792017 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gp5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1491f559-bc12-4afd-a40c-4eaa40d920a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gp5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:53Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.804911 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f8cd5bd-4a7f-4a5b-ab00-0cc64de9de94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cba320f443ed90a4e755228b663a1da3096e7ea8e068671d1f6273544b6ef60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c507106f39ae2e8acd4268c4e8f28ffd88a0965e69544d8d8a8830e83eceaf9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d62fbefb67a24fedc5f13652596c76f5b288fec58d79b5209f49c79db139d79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61317d26f40a4361a289779bb97b235525b8dddae383f63abfdcab634b633479\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\"\\\\nI0930 13:56:34.911386 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 13:56:34.920195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 13:56:34.920220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 13:56:34.920240 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 13:56:34.920248 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 13:56:34.927953 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 13:56:34.927977 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927984 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 13:56:34.927992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 13:56:34.927996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 13:56:34.928001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 13:56:34.928159 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0930 13:56:34.930945 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759240579\\\\\\\\\\\\\\\" (2025-09-30 13:56:19 +0000 UTC to 2025-10-30 13:56:20 +0000 UTC (now=2025-09-30 13:56:34.930900435 +0000 UTC))\\\\\\\"\\\\nF0930 13:56:34.930995 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1435e2e88d45e5561ff5863095d0f953ef1478d19d87691ab226cad7ae148f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:53Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.816977 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:53Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.820985 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.821026 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.821036 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.821051 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.821062 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:53Z","lastTransitionTime":"2025-09-30T13:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.835806 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84052a424fd6279d70b54902ebdcf430c8aa53b02406b52fc261db9d4d88ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27a94b18137fa991459484406b1c5bd2b2dbb52f527d5f13156ba8e5b8d0146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc4e5d3625aa2d5f94307cedddf72cc48f17f327a044501144e35422d670d555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9549ce85242f1f19b3a52e603caec58bb4c3e2449eb50d8c4525f56dfc93e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a364a13c773fe00ee4e022a286821013d3a681acc16567e66b1c33c7e18ee3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2322a3823793c1dd16af892010cd522aeb59d3ccb359e34c603a3ae0181d32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8cb31464b02d3ac9f62610436fdaa00d009301848cf17a62ec1f7deeef08fbf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"tor.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 13:56:49.801872 6077 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 13:56:49.802099 6077 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0930 13:56:49.802145 6077 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0930 13:56:49.802229 6077 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0930 13:56:49.802238 6077 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0930 13:56:49.802249 6077 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 13:56:49.802255 6077 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0930 13:56:49.802283 6077 factory.go:656] Stopping watch factory\\\\nI0930 13:56:49.802277 6077 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0930 13:56:49.802297 6077 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0930 13:56:49.802309 6077 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 13:56:49.802319 6077 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0930 13:56:49.802326 6077 handler.go:208] Removed *v1.Node event handler 2\\\\nI0930 13:56:49.802333 6077 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a2322a3823793c1dd16af892010cd522aeb59d3ccb359e34c603a3ae0181d32\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"message\\\":\\\"rred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z]\\\\nI0930 13:56:52.715429 6266 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-hmwzh in node crc\\\\nI0930 13:56:52.715417 6266 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"ba175bbe-5cc4-47e6-a32d-57693e1320bd\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-k\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:53Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.923601 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.923852 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.923926 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.923999 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:53 crc kubenswrapper[4840]: I0930 13:56:53.924059 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:53Z","lastTransitionTime":"2025-09-30T13:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.026690 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.026959 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.027024 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.027101 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.027162 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:54Z","lastTransitionTime":"2025-09-30T13:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.116255 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:56:54 crc kubenswrapper[4840]: E0930 13:56:54.116424 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.129844 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.129892 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.129904 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.129920 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.129933 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:54Z","lastTransitionTime":"2025-09-30T13:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.232574 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.232626 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.232638 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.232660 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.232674 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:54Z","lastTransitionTime":"2025-09-30T13:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.335239 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.335763 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.335950 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.336101 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.336230 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:54Z","lastTransitionTime":"2025-09-30T13:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.411503 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kfx69_2ff48c28-d076-46e8-a93f-9630989f81e8/ovnkube-controller/1.log" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.415275 4840 scope.go:117] "RemoveContainer" containerID="2a2322a3823793c1dd16af892010cd522aeb59d3ccb359e34c603a3ae0181d32" Sep 30 13:56:54 crc kubenswrapper[4840]: E0930 13:56:54.415419 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-kfx69_openshift-ovn-kubernetes(2ff48c28-d076-46e8-a93f-9630989f81e8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.428488 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:54Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.438893 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.438953 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.438967 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.438986 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.438999 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:54Z","lastTransitionTime":"2025-09-30T13:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.446897 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:54Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.465762 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:54Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.477742 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pj2zd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4de996b8-bda8-4142-b5e0-04ed9ae1f327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7286e2d685d6573993e2fa000cb9e08d85c6aa820af507327be52ca50fe58c7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pj2zd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:54Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.493224 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:54Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.508831 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:54Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.520840 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c2d8050-9531-4413-9d18-14bf3be65acf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34a542c60726f160b423c3e546ff4769a4a7e4a2441b5a09099e4d8c688286c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b5244dfa04a1f1804ef64dc65030d5eaedf679849eec6a9f0731e6cf147f83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7nvs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:54Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.533478 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gp5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1491f559-bc12-4afd-a40c-4eaa40d920a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gp5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:54Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.541099 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.541138 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.541148 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.541162 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.541171 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:54Z","lastTransitionTime":"2025-09-30T13:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.551753 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f8cd5bd-4a7f-4a5b-ab00-0cc64de9de94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cba320f443ed90a4e755228b663a1da3096e7ea8e068671d1f6273544b6ef60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c507106f39ae2e8acd4268c4e8f28ffd88a0965e69544d8d8a8830e83eceaf9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d62fbefb67a24fedc5f13652596c76f5b288fec58d79b5209f49c79db139d79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61317d26f40a4361a289779bb97b235525b8dddae383f63abfdcab634b633479\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\"\\\\nI0930 13:56:34.911386 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 13:56:34.920195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 13:56:34.920220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 13:56:34.920240 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 13:56:34.920248 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 13:56:34.927953 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 13:56:34.927977 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927984 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 13:56:34.927992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 13:56:34.927996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 13:56:34.928001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 13:56:34.928159 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0930 13:56:34.930945 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759240579\\\\\\\\\\\\\\\" (2025-09-30 13:56:19 +0000 UTC to 2025-10-30 13:56:20 +0000 UTC (now=2025-09-30 13:56:34.930900435 +0000 UTC))\\\\\\\"\\\\nF0930 13:56:34.930995 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1435e2e88d45e5561ff5863095d0f953ef1478d19d87691ab226cad7ae148f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:54Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.565476 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:54Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.582749 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8d388c29baf0ed22d5c9240f2585e4ca133de265227af6c802a169da6650f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:54Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.594770 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b16edc8f720d4296d7fe50fe407f5983ab975818ec6954d1f0e738e5596624a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:54Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.608446 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:54Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.630265 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84052a424fd6279d70b54902ebdcf430c8aa53b02406b52fc261db9d4d88ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27a94b18137fa991459484406b1c5bd2b2dbb52f527d5f13156ba8e5b8d0146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc4e5d3625aa2d5f94307cedddf72cc48f17f327a044501144e35422d670d555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9549ce85242f1f19b3a52e603caec58bb4c3e2449eb50d8c4525f56dfc93e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a364a13c773fe00ee4e022a286821013d3a681acc16567e66b1c33c7e18ee3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2322a3823793c1dd16af892010cd522aeb59d3ccb359e34c603a3ae0181d32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a2322a3823793c1dd16af892010cd522aeb59d3ccb359e34c603a3ae0181d32\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"message\\\":\\\"rred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z]\\\\nI0930 13:56:52.715429 6266 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-hmwzh in node crc\\\\nI0930 13:56:52.715417 6266 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"ba175bbe-5cc4-47e6-a32d-57693e1320bd\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-k\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-kfx69_openshift-ovn-kubernetes(2ff48c28-d076-46e8-a93f-9630989f81e8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:54Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.644484 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.644568 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.644582 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.644603 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.644616 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:54Z","lastTransitionTime":"2025-09-30T13:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.647486 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:54Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.747690 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.747738 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.747749 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.747766 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.747778 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:54Z","lastTransitionTime":"2025-09-30T13:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.849508 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.849599 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:54 crc kubenswrapper[4840]: E0930 13:56:54.849616 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:57:10.849597083 +0000 UTC m=+59.478683506 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.849625 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.849644 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.849663 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.849675 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.849713 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.849741 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.849674 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:54Z","lastTransitionTime":"2025-09-30T13:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.849775 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:56:54 crc kubenswrapper[4840]: E0930 13:56:54.849786 4840 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 13:56:54 crc kubenswrapper[4840]: E0930 13:56:54.849867 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 13:57:10.849844229 +0000 UTC m=+59.478930652 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 13:56:54 crc kubenswrapper[4840]: E0930 13:56:54.849901 4840 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 13:56:54 crc kubenswrapper[4840]: E0930 13:56:54.849917 4840 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 13:56:54 crc kubenswrapper[4840]: E0930 13:56:54.849928 4840 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 13:56:54 crc kubenswrapper[4840]: E0930 13:56:54.849951 4840 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 13:56:54 crc kubenswrapper[4840]: E0930 13:56:54.849999 4840 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 13:56:54 crc kubenswrapper[4840]: E0930 13:56:54.850029 4840 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 13:56:54 crc kubenswrapper[4840]: E0930 13:56:54.850049 4840 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 13:56:54 crc kubenswrapper[4840]: E0930 13:56:54.849968 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-30 13:57:10.849960092 +0000 UTC m=+59.479046515 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 13:56:54 crc kubenswrapper[4840]: E0930 13:56:54.850116 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 13:57:10.850081166 +0000 UTC m=+59.479167639 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 13:56:54 crc kubenswrapper[4840]: E0930 13:56:54.850153 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-30 13:57:10.850137307 +0000 UTC m=+59.479223860 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.952011 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.952061 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.952073 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.952093 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:54 crc kubenswrapper[4840]: I0930 13:56:54.952106 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:54Z","lastTransitionTime":"2025-09-30T13:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.054589 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.054628 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.054638 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.054653 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.054665 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:55Z","lastTransitionTime":"2025-09-30T13:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.115439 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.115493 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.115542 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:56:55 crc kubenswrapper[4840]: E0930 13:56:55.115576 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:56:55 crc kubenswrapper[4840]: E0930 13:56:55.115656 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:56:55 crc kubenswrapper[4840]: E0930 13:56:55.115721 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.156421 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.156458 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.156467 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.156482 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.156491 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:55Z","lastTransitionTime":"2025-09-30T13:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.259264 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.259353 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.259376 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.259407 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.259428 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:55Z","lastTransitionTime":"2025-09-30T13:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.362131 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.362192 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.362206 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.362223 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.362235 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:55Z","lastTransitionTime":"2025-09-30T13:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.465107 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.465149 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.465160 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.465175 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.465186 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:55Z","lastTransitionTime":"2025-09-30T13:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.567317 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.567384 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.567397 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.567414 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.567425 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:55Z","lastTransitionTime":"2025-09-30T13:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.669995 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.670061 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.670075 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.670095 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.670109 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:55Z","lastTransitionTime":"2025-09-30T13:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.761320 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.761397 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.761410 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.761428 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.761447 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:55Z","lastTransitionTime":"2025-09-30T13:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:55 crc kubenswrapper[4840]: E0930 13:56:55.793113 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5\\\",\\\"systemUUID\\\":\\\"f41e0351-4eda-411a-9a99-43a1952f3d34\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:55Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.799211 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.799253 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.799265 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.799286 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.799298 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:55Z","lastTransitionTime":"2025-09-30T13:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:55 crc kubenswrapper[4840]: E0930 13:56:55.822594 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5\\\",\\\"systemUUID\\\":\\\"f41e0351-4eda-411a-9a99-43a1952f3d34\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:55Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.829656 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.829706 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.829720 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.829744 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.829759 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:55Z","lastTransitionTime":"2025-09-30T13:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:55 crc kubenswrapper[4840]: E0930 13:56:55.845640 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5\\\",\\\"systemUUID\\\":\\\"f41e0351-4eda-411a-9a99-43a1952f3d34\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:55Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.849308 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.849357 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.849369 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.849406 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.849420 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:55Z","lastTransitionTime":"2025-09-30T13:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:55 crc kubenswrapper[4840]: E0930 13:56:55.860751 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5\\\",\\\"systemUUID\\\":\\\"f41e0351-4eda-411a-9a99-43a1952f3d34\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:55Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.863935 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.863976 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.863985 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.864000 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.864009 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:55Z","lastTransitionTime":"2025-09-30T13:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:55 crc kubenswrapper[4840]: E0930 13:56:55.874495 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5\\\",\\\"systemUUID\\\":\\\"f41e0351-4eda-411a-9a99-43a1952f3d34\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:55Z is after 2025-08-24T17:21:41Z" Sep 30 13:56:55 crc kubenswrapper[4840]: E0930 13:56:55.874634 4840 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.875629 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.875654 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.875661 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.875673 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.875683 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:55Z","lastTransitionTime":"2025-09-30T13:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.978482 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.978602 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.978623 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.978648 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:55 crc kubenswrapper[4840]: I0930 13:56:55.978667 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:55Z","lastTransitionTime":"2025-09-30T13:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.080975 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.081127 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.081142 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.081160 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.081169 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:56Z","lastTransitionTime":"2025-09-30T13:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.116066 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:56:56 crc kubenswrapper[4840]: E0930 13:56:56.116288 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.183853 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.183913 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.183924 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.183943 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.183954 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:56Z","lastTransitionTime":"2025-09-30T13:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.287352 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.287416 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.287433 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.287457 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.287475 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:56Z","lastTransitionTime":"2025-09-30T13:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.390231 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.390290 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.390309 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.390330 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.390341 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:56Z","lastTransitionTime":"2025-09-30T13:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.493292 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.493361 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.493387 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.493416 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.493437 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:56Z","lastTransitionTime":"2025-09-30T13:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.596186 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.596237 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.596249 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.596269 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.596282 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:56Z","lastTransitionTime":"2025-09-30T13:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.699076 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.699135 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.699151 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.699174 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.699191 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:56Z","lastTransitionTime":"2025-09-30T13:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.802116 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.802221 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.802244 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.802274 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.802293 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:56Z","lastTransitionTime":"2025-09-30T13:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.904996 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.905084 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.905095 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.905109 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:56 crc kubenswrapper[4840]: I0930 13:56:56.905119 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:56Z","lastTransitionTime":"2025-09-30T13:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.007658 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.007722 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.007762 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.007794 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.007814 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:57Z","lastTransitionTime":"2025-09-30T13:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.110212 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.110265 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.110280 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.110302 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.110317 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:57Z","lastTransitionTime":"2025-09-30T13:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.115622 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.115655 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.115668 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:56:57 crc kubenswrapper[4840]: E0930 13:56:57.115763 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:56:57 crc kubenswrapper[4840]: E0930 13:56:57.115859 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:56:57 crc kubenswrapper[4840]: E0930 13:56:57.115961 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.212384 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.212426 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.212471 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.212493 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.212505 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:57Z","lastTransitionTime":"2025-09-30T13:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.315271 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.315361 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.315409 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.315437 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.315492 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:57Z","lastTransitionTime":"2025-09-30T13:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.418593 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.418631 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.418639 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.418654 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.418663 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:57Z","lastTransitionTime":"2025-09-30T13:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.521137 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.521183 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.521195 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.521213 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.521227 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:57Z","lastTransitionTime":"2025-09-30T13:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.576532 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1491f559-bc12-4afd-a40c-4eaa40d920a8-metrics-certs\") pod \"network-metrics-daemon-4gp5x\" (UID: \"1491f559-bc12-4afd-a40c-4eaa40d920a8\") " pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:56:57 crc kubenswrapper[4840]: E0930 13:56:57.576752 4840 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 13:56:57 crc kubenswrapper[4840]: E0930 13:56:57.576836 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1491f559-bc12-4afd-a40c-4eaa40d920a8-metrics-certs podName:1491f559-bc12-4afd-a40c-4eaa40d920a8 nodeName:}" failed. No retries permitted until 2025-09-30 13:57:05.576813713 +0000 UTC m=+54.205900136 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1491f559-bc12-4afd-a40c-4eaa40d920a8-metrics-certs") pod "network-metrics-daemon-4gp5x" (UID: "1491f559-bc12-4afd-a40c-4eaa40d920a8") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.623612 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.623672 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.623682 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.623696 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.623705 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:57Z","lastTransitionTime":"2025-09-30T13:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.726314 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.726365 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.726375 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.726392 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.726419 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:57Z","lastTransitionTime":"2025-09-30T13:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.828905 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.828950 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.828961 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.828978 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.828988 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:57Z","lastTransitionTime":"2025-09-30T13:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.932315 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.932387 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.932405 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.932428 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:57 crc kubenswrapper[4840]: I0930 13:56:57.932451 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:57Z","lastTransitionTime":"2025-09-30T13:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.035159 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.035195 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.035206 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.035221 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.035230 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:58Z","lastTransitionTime":"2025-09-30T13:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.115930 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:56:58 crc kubenswrapper[4840]: E0930 13:56:58.116079 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.136979 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.137023 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.137036 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.137052 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.137063 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:58Z","lastTransitionTime":"2025-09-30T13:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.239628 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.239668 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.239678 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.239692 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.239702 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:58Z","lastTransitionTime":"2025-09-30T13:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.342473 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.342526 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.342539 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.342569 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.342581 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:58Z","lastTransitionTime":"2025-09-30T13:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.444863 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.444915 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.444926 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.444942 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.444953 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:58Z","lastTransitionTime":"2025-09-30T13:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.549086 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.549169 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.549282 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.549309 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.549319 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:58Z","lastTransitionTime":"2025-09-30T13:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.652198 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.652241 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.652252 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.652268 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.652280 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:58Z","lastTransitionTime":"2025-09-30T13:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.754260 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.754301 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.754312 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.754324 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.754333 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:58Z","lastTransitionTime":"2025-09-30T13:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.857326 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.857363 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.857375 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.857391 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.857403 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:58Z","lastTransitionTime":"2025-09-30T13:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.959639 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.959688 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.959708 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.959731 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:58 crc kubenswrapper[4840]: I0930 13:56:58.959745 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:58Z","lastTransitionTime":"2025-09-30T13:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.062390 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.062444 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.062453 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.062465 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.062473 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:59Z","lastTransitionTime":"2025-09-30T13:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.115344 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.115452 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.115582 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:56:59 crc kubenswrapper[4840]: E0930 13:56:59.115478 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:56:59 crc kubenswrapper[4840]: E0930 13:56:59.115770 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:56:59 crc kubenswrapper[4840]: E0930 13:56:59.115991 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.165262 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.165309 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.165354 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.165375 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.165390 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:59Z","lastTransitionTime":"2025-09-30T13:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.268427 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.268511 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.268539 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.268632 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.268665 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:59Z","lastTransitionTime":"2025-09-30T13:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.370393 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.370439 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.370447 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.370465 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.370474 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:59Z","lastTransitionTime":"2025-09-30T13:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.474014 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.474096 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.474110 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.474127 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.474139 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:59Z","lastTransitionTime":"2025-09-30T13:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.577801 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.577840 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.577849 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.577862 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.577871 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:59Z","lastTransitionTime":"2025-09-30T13:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.679791 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.679837 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.679848 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.679863 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.679875 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:59Z","lastTransitionTime":"2025-09-30T13:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.782901 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.782945 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.782963 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.782988 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.783005 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:59Z","lastTransitionTime":"2025-09-30T13:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.886331 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.886395 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.886415 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.886461 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.886491 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:59Z","lastTransitionTime":"2025-09-30T13:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.989124 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.989189 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.989209 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.989249 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:56:59 crc kubenswrapper[4840]: I0930 13:56:59.989287 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:56:59Z","lastTransitionTime":"2025-09-30T13:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.092202 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.092246 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.092257 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.092276 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.092291 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:00Z","lastTransitionTime":"2025-09-30T13:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.116088 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:57:00 crc kubenswrapper[4840]: E0930 13:57:00.116243 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.194404 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.194454 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.194480 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.194497 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.194519 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:00Z","lastTransitionTime":"2025-09-30T13:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.297480 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.297570 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.297583 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.297602 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.297615 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:00Z","lastTransitionTime":"2025-09-30T13:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.400673 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.400728 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.400737 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.400753 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.400766 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:00Z","lastTransitionTime":"2025-09-30T13:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.502418 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.502454 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.502468 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.502486 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.502496 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:00Z","lastTransitionTime":"2025-09-30T13:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.604181 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.604238 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.604253 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.604275 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.604286 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:00Z","lastTransitionTime":"2025-09-30T13:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.706709 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.706745 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.706758 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.706775 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.706787 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:00Z","lastTransitionTime":"2025-09-30T13:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.809430 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.809469 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.809479 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.809494 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.809505 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:00Z","lastTransitionTime":"2025-09-30T13:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.913892 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.913941 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.913951 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.913966 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:00 crc kubenswrapper[4840]: I0930 13:57:00.913978 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:00Z","lastTransitionTime":"2025-09-30T13:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.016384 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.016438 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.016451 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.016473 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.016485 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:01Z","lastTransitionTime":"2025-09-30T13:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.116261 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.116261 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:57:01 crc kubenswrapper[4840]: E0930 13:57:01.116456 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.116296 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:57:01 crc kubenswrapper[4840]: E0930 13:57:01.116585 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:57:01 crc kubenswrapper[4840]: E0930 13:57:01.116767 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.119197 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.119230 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.119243 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.119258 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.119269 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:01Z","lastTransitionTime":"2025-09-30T13:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.221826 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.221876 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.221885 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.221900 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.221909 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:01Z","lastTransitionTime":"2025-09-30T13:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.324377 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.324432 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.324446 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.324480 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.324496 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:01Z","lastTransitionTime":"2025-09-30T13:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.427321 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.427364 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.427373 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.427395 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.427405 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:01Z","lastTransitionTime":"2025-09-30T13:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.529352 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.529418 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.529437 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.529461 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.529478 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:01Z","lastTransitionTime":"2025-09-30T13:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.631491 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.631529 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.631537 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.631570 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.631582 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:01Z","lastTransitionTime":"2025-09-30T13:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.733511 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.733567 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.733581 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.733602 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.733613 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:01Z","lastTransitionTime":"2025-09-30T13:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.836368 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.836437 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.836460 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.836489 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.836509 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:01Z","lastTransitionTime":"2025-09-30T13:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.938883 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.938928 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.938937 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.938952 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:01 crc kubenswrapper[4840]: I0930 13:57:01.938962 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:01Z","lastTransitionTime":"2025-09-30T13:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.042027 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.042097 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.042112 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.042139 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.042159 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:02Z","lastTransitionTime":"2025-09-30T13:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.115762 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:57:02 crc kubenswrapper[4840]: E0930 13:57:02.116040 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.142534 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84052a424fd6279d70b54902ebdcf430c8aa53b02406b52fc261db9d4d88ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27a94b18137fa991459484406b1c5bd2b2dbb52f527d5f13156ba8e5b8d0146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc4e5d3625aa2d5f94307cedddf72cc48f17f327a044501144e35422d670d555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9549ce85242f1f19b3a52e603caec58bb4c3e2449eb50d8c4525f56dfc93e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a364a13c773fe00ee4e022a286821013d3a681acc16567e66b1c33c7e18ee3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2322a3823793c1dd16af892010cd522aeb59d3ccb359e34c603a3ae0181d32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a2322a3823793c1dd16af892010cd522aeb59d3ccb359e34c603a3ae0181d32\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"message\\\":\\\"rred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z]\\\\nI0930 13:56:52.715429 6266 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-hmwzh in node crc\\\\nI0930 13:56:52.715417 6266 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"ba175bbe-5cc4-47e6-a32d-57693e1320bd\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-k\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-kfx69_openshift-ovn-kubernetes(2ff48c28-d076-46e8-a93f-9630989f81e8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:02Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.145435 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.145485 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.145508 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.145537 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.145592 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:02Z","lastTransitionTime":"2025-09-30T13:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.160600 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:02Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.172749 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:02Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.182756 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pj2zd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4de996b8-bda8-4142-b5e0-04ed9ae1f327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7286e2d685d6573993e2fa000cb9e08d85c6aa820af507327be52ca50fe58c7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pj2zd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:02Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.196336 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:02Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.207735 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:02Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.219917 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:02Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.234138 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:02Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.247100 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.247258 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.247266 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.247278 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.247288 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:02Z","lastTransitionTime":"2025-09-30T13:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.258471 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8d388c29baf0ed22d5c9240f2585e4ca133de265227af6c802a169da6650f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:02Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.269528 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b16edc8f720d4296d7fe50fe407f5983ab975818ec6954d1f0e738e5596624a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:02Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.282996 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:02Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.294065 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c2d8050-9531-4413-9d18-14bf3be65acf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34a542c60726f160b423c3e546ff4769a4a7e4a2441b5a09099e4d8c688286c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b5244dfa04a1f1804ef64dc65030d5eaedf679849eec6a9f0731e6cf147f83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7nvs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:02Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.304810 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gp5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1491f559-bc12-4afd-a40c-4eaa40d920a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gp5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:02Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.318581 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f8cd5bd-4a7f-4a5b-ab00-0cc64de9de94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cba320f443ed90a4e755228b663a1da3096e7ea8e068671d1f6273544b6ef60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c507106f39ae2e8acd4268c4e8f28ffd88a0965e69544d8d8a8830e83eceaf9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d62fbefb67a24fedc5f13652596c76f5b288fec58d79b5209f49c79db139d79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61317d26f40a4361a289779bb97b235525b8dddae383f63abfdcab634b633479\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\"\\\\nI0930 13:56:34.911386 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 13:56:34.920195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 13:56:34.920220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 13:56:34.920240 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 13:56:34.920248 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 13:56:34.927953 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 13:56:34.927977 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927984 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 13:56:34.927992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 13:56:34.927996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 13:56:34.928001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 13:56:34.928159 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0930 13:56:34.930945 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759240579\\\\\\\\\\\\\\\" (2025-09-30 13:56:19 +0000 UTC to 2025-10-30 13:56:20 +0000 UTC (now=2025-09-30 13:56:34.930900435 +0000 UTC))\\\\\\\"\\\\nF0930 13:56:34.930995 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1435e2e88d45e5561ff5863095d0f953ef1478d19d87691ab226cad7ae148f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:02Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.334183 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:02Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.349089 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.349115 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.349124 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.349138 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.349148 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:02Z","lastTransitionTime":"2025-09-30T13:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.451481 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.451529 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.451540 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.451574 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.451586 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:02Z","lastTransitionTime":"2025-09-30T13:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.554208 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.554252 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.554266 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.554283 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.554293 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:02Z","lastTransitionTime":"2025-09-30T13:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.656999 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.657100 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.657126 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.657164 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.657191 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:02Z","lastTransitionTime":"2025-09-30T13:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.759496 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.759568 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.759579 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.759596 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.759606 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:02Z","lastTransitionTime":"2025-09-30T13:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.862847 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.862930 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.862959 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.862991 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.863018 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:02Z","lastTransitionTime":"2025-09-30T13:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.965613 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.965665 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.965675 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.965692 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:02 crc kubenswrapper[4840]: I0930 13:57:02.965703 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:02Z","lastTransitionTime":"2025-09-30T13:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.068533 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.068616 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.068634 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.068656 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.068672 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:03Z","lastTransitionTime":"2025-09-30T13:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.116176 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.116227 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.116270 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:57:03 crc kubenswrapper[4840]: E0930 13:57:03.116347 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:57:03 crc kubenswrapper[4840]: E0930 13:57:03.116476 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:57:03 crc kubenswrapper[4840]: E0930 13:57:03.116662 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.171499 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.171537 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.171545 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.171577 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.171587 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:03Z","lastTransitionTime":"2025-09-30T13:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.274164 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.274200 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.274212 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.274225 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.274234 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:03Z","lastTransitionTime":"2025-09-30T13:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.377274 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.377316 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.377329 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.377346 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.377362 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:03Z","lastTransitionTime":"2025-09-30T13:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.480189 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.480260 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.480280 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.480298 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.480310 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:03Z","lastTransitionTime":"2025-09-30T13:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.582974 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.583041 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.583062 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.583087 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.583105 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:03Z","lastTransitionTime":"2025-09-30T13:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.685266 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.685332 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.685351 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.685376 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.685394 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:03Z","lastTransitionTime":"2025-09-30T13:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.788008 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.788058 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.788070 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.788086 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.788099 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:03Z","lastTransitionTime":"2025-09-30T13:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.890385 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.890417 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.890427 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.890442 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.890451 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:03Z","lastTransitionTime":"2025-09-30T13:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.993130 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.993189 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.993201 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.993217 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:03 crc kubenswrapper[4840]: I0930 13:57:03.993229 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:03Z","lastTransitionTime":"2025-09-30T13:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.096903 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.096986 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.097008 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.097037 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.097059 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:04Z","lastTransitionTime":"2025-09-30T13:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.116284 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:57:04 crc kubenswrapper[4840]: E0930 13:57:04.116464 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.200648 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.200734 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.200757 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.200793 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.200818 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:04Z","lastTransitionTime":"2025-09-30T13:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.304054 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.304134 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.304153 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.304186 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.304212 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:04Z","lastTransitionTime":"2025-09-30T13:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.406494 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.406527 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.406535 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.406564 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.406574 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:04Z","lastTransitionTime":"2025-09-30T13:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.509776 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.509835 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.509848 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.509871 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.509888 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:04Z","lastTransitionTime":"2025-09-30T13:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.612767 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.612847 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.612866 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.612893 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.612979 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:04Z","lastTransitionTime":"2025-09-30T13:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.715444 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.715526 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.715547 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.715609 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.715630 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:04Z","lastTransitionTime":"2025-09-30T13:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.818806 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.818848 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.818857 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.818872 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.818881 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:04Z","lastTransitionTime":"2025-09-30T13:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.921806 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.921870 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.921882 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.921899 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:04 crc kubenswrapper[4840]: I0930 13:57:04.921912 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:04Z","lastTransitionTime":"2025-09-30T13:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.024171 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.024213 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.024227 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.024246 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.024258 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:05Z","lastTransitionTime":"2025-09-30T13:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.115955 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.116051 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.116052 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:57:05 crc kubenswrapper[4840]: E0930 13:57:05.116160 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:57:05 crc kubenswrapper[4840]: E0930 13:57:05.116272 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:57:05 crc kubenswrapper[4840]: E0930 13:57:05.116333 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.126282 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.126313 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.126322 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.126338 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.126346 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:05Z","lastTransitionTime":"2025-09-30T13:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.228325 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.228390 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.228410 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.228436 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.228456 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:05Z","lastTransitionTime":"2025-09-30T13:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.331256 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.331303 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.331314 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.331332 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.331344 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:05Z","lastTransitionTime":"2025-09-30T13:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.433359 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.433444 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.433469 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.433501 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.433525 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:05Z","lastTransitionTime":"2025-09-30T13:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.535906 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.535943 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.535954 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.535969 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.535979 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:05Z","lastTransitionTime":"2025-09-30T13:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.638037 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.638096 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.638115 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.638132 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.638144 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:05Z","lastTransitionTime":"2025-09-30T13:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.659760 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1491f559-bc12-4afd-a40c-4eaa40d920a8-metrics-certs\") pod \"network-metrics-daemon-4gp5x\" (UID: \"1491f559-bc12-4afd-a40c-4eaa40d920a8\") " pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:57:05 crc kubenswrapper[4840]: E0930 13:57:05.659888 4840 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 13:57:05 crc kubenswrapper[4840]: E0930 13:57:05.659945 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1491f559-bc12-4afd-a40c-4eaa40d920a8-metrics-certs podName:1491f559-bc12-4afd-a40c-4eaa40d920a8 nodeName:}" failed. No retries permitted until 2025-09-30 13:57:21.659929475 +0000 UTC m=+70.289015898 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1491f559-bc12-4afd-a40c-4eaa40d920a8-metrics-certs") pod "network-metrics-daemon-4gp5x" (UID: "1491f559-bc12-4afd-a40c-4eaa40d920a8") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.740864 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.740919 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.740931 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.740950 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.740963 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:05Z","lastTransitionTime":"2025-09-30T13:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.842872 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.842954 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.842963 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.842978 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.842986 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:05Z","lastTransitionTime":"2025-09-30T13:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.945546 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.945620 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.945631 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.945649 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:05 crc kubenswrapper[4840]: I0930 13:57:05.945662 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:05Z","lastTransitionTime":"2025-09-30T13:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.045798 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.045836 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.045847 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.045864 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.045876 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:06Z","lastTransitionTime":"2025-09-30T13:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:06 crc kubenswrapper[4840]: E0930 13:57:06.062861 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5\\\",\\\"systemUUID\\\":\\\"f41e0351-4eda-411a-9a99-43a1952f3d34\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:06Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.066164 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.066184 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.066192 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.066218 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.066229 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:06Z","lastTransitionTime":"2025-09-30T13:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:06 crc kubenswrapper[4840]: E0930 13:57:06.080701 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5\\\",\\\"systemUUID\\\":\\\"f41e0351-4eda-411a-9a99-43a1952f3d34\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:06Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.084157 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.084183 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.084194 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.084211 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.084221 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:06Z","lastTransitionTime":"2025-09-30T13:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:06 crc kubenswrapper[4840]: E0930 13:57:06.101325 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5\\\",\\\"systemUUID\\\":\\\"f41e0351-4eda-411a-9a99-43a1952f3d34\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:06Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.105292 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.105326 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.105334 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.105347 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.105357 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:06Z","lastTransitionTime":"2025-09-30T13:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.116065 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:57:06 crc kubenswrapper[4840]: E0930 13:57:06.116257 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:57:06 crc kubenswrapper[4840]: E0930 13:57:06.119013 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5\\\",\\\"systemUUID\\\":\\\"f41e0351-4eda-411a-9a99-43a1952f3d34\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:06Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.123540 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.123588 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.123596 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.123614 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.123630 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:06Z","lastTransitionTime":"2025-09-30T13:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:06 crc kubenswrapper[4840]: E0930 13:57:06.141058 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5\\\",\\\"systemUUID\\\":\\\"f41e0351-4eda-411a-9a99-43a1952f3d34\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:06Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:06 crc kubenswrapper[4840]: E0930 13:57:06.141287 4840 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.142799 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.142845 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.142862 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.142885 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.142901 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:06Z","lastTransitionTime":"2025-09-30T13:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.244947 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.244997 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.245005 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.245020 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.245029 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:06Z","lastTransitionTime":"2025-09-30T13:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.348068 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.348117 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.348130 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.348150 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.348164 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:06Z","lastTransitionTime":"2025-09-30T13:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.451475 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.451590 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.451611 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.451637 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.451655 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:06Z","lastTransitionTime":"2025-09-30T13:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.555145 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.555202 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.555222 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.555249 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.555268 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:06Z","lastTransitionTime":"2025-09-30T13:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.658935 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.658989 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.658999 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.659018 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.659031 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:06Z","lastTransitionTime":"2025-09-30T13:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.761421 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.761482 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.761503 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.761530 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.761590 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:06Z","lastTransitionTime":"2025-09-30T13:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.864012 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.864056 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.864071 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.864088 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.864098 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:06Z","lastTransitionTime":"2025-09-30T13:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.966402 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.966444 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.966453 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.966468 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:06 crc kubenswrapper[4840]: I0930 13:57:06.966478 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:06Z","lastTransitionTime":"2025-09-30T13:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.032246 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.043270 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.045682 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:07Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.060149 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:07Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.068458 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.068486 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.068498 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.068513 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.068524 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:07Z","lastTransitionTime":"2025-09-30T13:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.074689 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:07Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.090966 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8d388c29baf0ed22d5c9240f2585e4ca133de265227af6c802a169da6650f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:07Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.102742 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b16edc8f720d4296d7fe50fe407f5983ab975818ec6954d1f0e738e5596624a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:07Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.112308 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:07Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.115690 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.115737 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.115690 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:57:07 crc kubenswrapper[4840]: E0930 13:57:07.115816 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:57:07 crc kubenswrapper[4840]: E0930 13:57:07.115932 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:57:07 crc kubenswrapper[4840]: E0930 13:57:07.116081 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.124590 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c2d8050-9531-4413-9d18-14bf3be65acf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34a542c60726f160b423c3e546ff4769a4a7e4a2441b5a09099e4d8c688286c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b5244dfa04a1f1804ef64dc65030d5eaedf679849eec6a9f0731e6cf147f83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7nvs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:07Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.136399 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gp5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1491f559-bc12-4afd-a40c-4eaa40d920a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gp5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:07Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.151505 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f8cd5bd-4a7f-4a5b-ab00-0cc64de9de94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cba320f443ed90a4e755228b663a1da3096e7ea8e068671d1f6273544b6ef60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c507106f39ae2e8acd4268c4e8f28ffd88a0965e69544d8d8a8830e83eceaf9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d62fbefb67a24fedc5f13652596c76f5b288fec58d79b5209f49c79db139d79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61317d26f40a4361a289779bb97b235525b8dddae383f63abfdcab634b633479\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\"\\\\nI0930 13:56:34.911386 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 13:56:34.920195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 13:56:34.920220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 13:56:34.920240 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 13:56:34.920248 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 13:56:34.927953 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 13:56:34.927977 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927984 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 13:56:34.927992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 13:56:34.927996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 13:56:34.928001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 13:56:34.928159 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0930 13:56:34.930945 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759240579\\\\\\\\\\\\\\\" (2025-09-30 13:56:19 +0000 UTC to 2025-10-30 13:56:20 +0000 UTC (now=2025-09-30 13:56:34.930900435 +0000 UTC))\\\\\\\"\\\\nF0930 13:56:34.930995 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1435e2e88d45e5561ff5863095d0f953ef1478d19d87691ab226cad7ae148f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:07Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.163502 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:07Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.170166 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.170202 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.170215 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.170231 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.170244 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:07Z","lastTransitionTime":"2025-09-30T13:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.190176 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84052a424fd6279d70b54902ebdcf430c8aa53b02406b52fc261db9d4d88ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27a94b18137fa991459484406b1c5bd2b2dbb52f527d5f13156ba8e5b8d0146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc4e5d3625aa2d5f94307cedddf72cc48f17f327a044501144e35422d670d555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9549ce85242f1f19b3a52e603caec58bb4c3e2449eb50d8c4525f56dfc93e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a364a13c773fe00ee4e022a286821013d3a681acc16567e66b1c33c7e18ee3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2322a3823793c1dd16af892010cd522aeb59d3ccb359e34c603a3ae0181d32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a2322a3823793c1dd16af892010cd522aeb59d3ccb359e34c603a3ae0181d32\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"message\\\":\\\"rred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z]\\\\nI0930 13:56:52.715429 6266 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-hmwzh in node crc\\\\nI0930 13:56:52.715417 6266 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"ba175bbe-5cc4-47e6-a32d-57693e1320bd\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-k\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-kfx69_openshift-ovn-kubernetes(2ff48c28-d076-46e8-a93f-9630989f81e8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:07Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.203697 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:07Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.216698 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:07Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.228098 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pj2zd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4de996b8-bda8-4142-b5e0-04ed9ae1f327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7286e2d685d6573993e2fa000cb9e08d85c6aa820af507327be52ca50fe58c7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pj2zd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:07Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.240431 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:07Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.272850 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.272893 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.272903 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.272920 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.272932 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:07Z","lastTransitionTime":"2025-09-30T13:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.377872 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.377900 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.377913 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.377926 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.377936 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:07Z","lastTransitionTime":"2025-09-30T13:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.479605 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.479655 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.479667 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.479680 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.479689 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:07Z","lastTransitionTime":"2025-09-30T13:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.581761 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.581790 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.581800 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.581817 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.581828 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:07Z","lastTransitionTime":"2025-09-30T13:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.683956 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.684003 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.684011 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.684025 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.684034 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:07Z","lastTransitionTime":"2025-09-30T13:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.785835 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.785877 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.785887 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.785900 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.785908 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:07Z","lastTransitionTime":"2025-09-30T13:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.887859 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.887897 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.887907 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.887921 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.887931 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:07Z","lastTransitionTime":"2025-09-30T13:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.990012 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.990059 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.990071 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.990086 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:07 crc kubenswrapper[4840]: I0930 13:57:07.990098 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:07Z","lastTransitionTime":"2025-09-30T13:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.092577 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.092620 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.092631 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.092651 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.092664 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:08Z","lastTransitionTime":"2025-09-30T13:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.116045 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:57:08 crc kubenswrapper[4840]: E0930 13:57:08.116179 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.117344 4840 scope.go:117] "RemoveContainer" containerID="2a2322a3823793c1dd16af892010cd522aeb59d3ccb359e34c603a3ae0181d32" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.195994 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.196439 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.196449 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.196465 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.196478 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:08Z","lastTransitionTime":"2025-09-30T13:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.298357 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.298389 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.298397 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.298410 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.298420 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:08Z","lastTransitionTime":"2025-09-30T13:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.401002 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.401037 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.401046 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.401061 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.401070 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:08Z","lastTransitionTime":"2025-09-30T13:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.460893 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kfx69_2ff48c28-d076-46e8-a93f-9630989f81e8/ovnkube-controller/1.log" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.463702 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" event={"ID":"2ff48c28-d076-46e8-a93f-9630989f81e8","Type":"ContainerStarted","Data":"c79f4e921ab2d7e8370fe83552c960147cbbd744f03bca3e390813ca3d021aa3"} Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.464286 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.478726 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:08Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.493915 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:08Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.503752 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.504027 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.504123 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.504204 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.504264 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:08Z","lastTransitionTime":"2025-09-30T13:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.507495 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gp5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1491f559-bc12-4afd-a40c-4eaa40d920a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gp5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:08Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.525884 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49c5addb-0d04-4c3d-b10e-a07d5fec55da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cf852cfa87ef328d172369c26197a9e6a58fe26fa3113469d9275b2e78c30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a08b53dbed5f3a033d2838519733a2a4109ac5a954d9f5fabd0762b2823b090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3f57df6d22c47163751c297aff0ca1dc2bdbf150b9e385e72acdefde92910c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08be626d75cb6bf53342c4ccab7abef9374735cd29d4afb7c7b5b62715db5d22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08be626d75cb6bf53342c4ccab7abef9374735cd29d4afb7c7b5b62715db5d22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:08Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.543278 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f8cd5bd-4a7f-4a5b-ab00-0cc64de9de94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cba320f443ed90a4e755228b663a1da3096e7ea8e068671d1f6273544b6ef60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c507106f39ae2e8acd4268c4e8f28ffd88a0965e69544d8d8a8830e83eceaf9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d62fbefb67a24fedc5f13652596c76f5b288fec58d79b5209f49c79db139d79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61317d26f40a4361a289779bb97b235525b8dddae383f63abfdcab634b633479\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\"\\\\nI0930 13:56:34.911386 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 13:56:34.920195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 13:56:34.920220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 13:56:34.920240 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 13:56:34.920248 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 13:56:34.927953 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 13:56:34.927977 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927984 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 13:56:34.927992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 13:56:34.927996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 13:56:34.928001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 13:56:34.928159 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0930 13:56:34.930945 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759240579\\\\\\\\\\\\\\\" (2025-09-30 13:56:19 +0000 UTC to 2025-10-30 13:56:20 +0000 UTC (now=2025-09-30 13:56:34.930900435 +0000 UTC))\\\\\\\"\\\\nF0930 13:56:34.930995 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1435e2e88d45e5561ff5863095d0f953ef1478d19d87691ab226cad7ae148f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:08Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.556440 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:08Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.574937 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8d388c29baf0ed22d5c9240f2585e4ca133de265227af6c802a169da6650f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:08Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.597206 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b16edc8f720d4296d7fe50fe407f5983ab975818ec6954d1f0e738e5596624a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:08Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.606207 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.606238 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.606246 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.606259 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.606268 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:08Z","lastTransitionTime":"2025-09-30T13:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.612391 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:08Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.629069 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c2d8050-9531-4413-9d18-14bf3be65acf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34a542c60726f160b423c3e546ff4769a4a7e4a2441b5a09099e4d8c688286c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b5244dfa04a1f1804ef64dc65030d5eaedf679849eec6a9f0731e6cf147f83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7nvs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:08Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.651501 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84052a424fd6279d70b54902ebdcf430c8aa53b02406b52fc261db9d4d88ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27a94b18137fa991459484406b1c5bd2b2dbb52f527d5f13156ba8e5b8d0146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc4e5d3625aa2d5f94307cedddf72cc48f17f327a044501144e35422d670d555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9549ce85242f1f19b3a52e603caec58bb4c3e2449eb50d8c4525f56dfc93e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a364a13c773fe00ee4e022a286821013d3a681acc16567e66b1c33c7e18ee3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79f4e921ab2d7e8370fe83552c960147cbbd744f03bca3e390813ca3d021aa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a2322a3823793c1dd16af892010cd522aeb59d3ccb359e34c603a3ae0181d32\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"message\\\":\\\"rred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z]\\\\nI0930 13:56:52.715429 6266 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-hmwzh in node crc\\\\nI0930 13:56:52.715417 6266 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"ba175bbe-5cc4-47e6-a32d-57693e1320bd\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-k\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:08Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.667980 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:08Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.681715 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:08Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.693269 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:08Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.708121 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.708161 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.708172 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.708189 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.708204 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:08Z","lastTransitionTime":"2025-09-30T13:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.711096 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:08Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.730391 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pj2zd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4de996b8-bda8-4142-b5e0-04ed9ae1f327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7286e2d685d6573993e2fa000cb9e08d85c6aa820af507327be52ca50fe58c7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pj2zd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:08Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.810165 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.810212 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.810223 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.810241 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.810252 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:08Z","lastTransitionTime":"2025-09-30T13:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.912605 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.912650 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.912660 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.912678 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:08 crc kubenswrapper[4840]: I0930 13:57:08.912690 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:08Z","lastTransitionTime":"2025-09-30T13:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.014851 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.014894 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.014903 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.014918 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.014927 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:09Z","lastTransitionTime":"2025-09-30T13:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.115884 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.115943 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.115954 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:57:09 crc kubenswrapper[4840]: E0930 13:57:09.116018 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:57:09 crc kubenswrapper[4840]: E0930 13:57:09.116203 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:57:09 crc kubenswrapper[4840]: E0930 13:57:09.116314 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.117163 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.117191 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.117202 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.117219 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.117231 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:09Z","lastTransitionTime":"2025-09-30T13:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.218937 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.218985 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.218996 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.219015 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.219027 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:09Z","lastTransitionTime":"2025-09-30T13:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.321744 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.321788 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.321801 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.321819 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.321830 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:09Z","lastTransitionTime":"2025-09-30T13:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.424323 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.424399 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.424422 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.424450 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.424471 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:09Z","lastTransitionTime":"2025-09-30T13:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.468820 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kfx69_2ff48c28-d076-46e8-a93f-9630989f81e8/ovnkube-controller/2.log" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.469671 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kfx69_2ff48c28-d076-46e8-a93f-9630989f81e8/ovnkube-controller/1.log" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.472597 4840 generic.go:334] "Generic (PLEG): container finished" podID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerID="c79f4e921ab2d7e8370fe83552c960147cbbd744f03bca3e390813ca3d021aa3" exitCode=1 Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.472662 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" event={"ID":"2ff48c28-d076-46e8-a93f-9630989f81e8","Type":"ContainerDied","Data":"c79f4e921ab2d7e8370fe83552c960147cbbd744f03bca3e390813ca3d021aa3"} Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.472712 4840 scope.go:117] "RemoveContainer" containerID="2a2322a3823793c1dd16af892010cd522aeb59d3ccb359e34c603a3ae0181d32" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.473443 4840 scope.go:117] "RemoveContainer" containerID="c79f4e921ab2d7e8370fe83552c960147cbbd744f03bca3e390813ca3d021aa3" Sep 30 13:57:09 crc kubenswrapper[4840]: E0930 13:57:09.473652 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-kfx69_openshift-ovn-kubernetes(2ff48c28-d076-46e8-a93f-9630989f81e8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.488184 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c2d8050-9531-4413-9d18-14bf3be65acf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34a542c60726f160b423c3e546ff4769a4a7e4a2441b5a09099e4d8c688286c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b5244dfa04a1f1804ef64dc65030d5eaedf679849eec6a9f0731e6cf147f83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7nvs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:09Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.500055 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gp5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1491f559-bc12-4afd-a40c-4eaa40d920a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gp5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:09Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.512272 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49c5addb-0d04-4c3d-b10e-a07d5fec55da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cf852cfa87ef328d172369c26197a9e6a58fe26fa3113469d9275b2e78c30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a08b53dbed5f3a033d2838519733a2a4109ac5a954d9f5fabd0762b2823b090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3f57df6d22c47163751c297aff0ca1dc2bdbf150b9e385e72acdefde92910c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08be626d75cb6bf53342c4ccab7abef9374735cd29d4afb7c7b5b62715db5d22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08be626d75cb6bf53342c4ccab7abef9374735cd29d4afb7c7b5b62715db5d22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:09Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.525192 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f8cd5bd-4a7f-4a5b-ab00-0cc64de9de94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cba320f443ed90a4e755228b663a1da3096e7ea8e068671d1f6273544b6ef60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c507106f39ae2e8acd4268c4e8f28ffd88a0965e69544d8d8a8830e83eceaf9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d62fbefb67a24fedc5f13652596c76f5b288fec58d79b5209f49c79db139d79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61317d26f40a4361a289779bb97b235525b8dddae383f63abfdcab634b633479\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\"\\\\nI0930 13:56:34.911386 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 13:56:34.920195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 13:56:34.920220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 13:56:34.920240 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 13:56:34.920248 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 13:56:34.927953 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 13:56:34.927977 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927984 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 13:56:34.927992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 13:56:34.927996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 13:56:34.928001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 13:56:34.928159 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0930 13:56:34.930945 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759240579\\\\\\\\\\\\\\\" (2025-09-30 13:56:19 +0000 UTC to 2025-10-30 13:56:20 +0000 UTC (now=2025-09-30 13:56:34.930900435 +0000 UTC))\\\\\\\"\\\\nF0930 13:56:34.930995 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1435e2e88d45e5561ff5863095d0f953ef1478d19d87691ab226cad7ae148f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:09Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.526524 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.526590 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.526603 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.526622 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.526640 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:09Z","lastTransitionTime":"2025-09-30T13:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.539843 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:09Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.556438 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8d388c29baf0ed22d5c9240f2585e4ca133de265227af6c802a169da6650f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:09Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.570200 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b16edc8f720d4296d7fe50fe407f5983ab975818ec6954d1f0e738e5596624a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:09Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.582356 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:09Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.601622 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84052a424fd6279d70b54902ebdcf430c8aa53b02406b52fc261db9d4d88ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27a94b18137fa991459484406b1c5bd2b2dbb52f527d5f13156ba8e5b8d0146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc4e5d3625aa2d5f94307cedddf72cc48f17f327a044501144e35422d670d555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9549ce85242f1f19b3a52e603caec58bb4c3e2449eb50d8c4525f56dfc93e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a364a13c773fe00ee4e022a286821013d3a681acc16567e66b1c33c7e18ee3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79f4e921ab2d7e8370fe83552c960147cbbd744f03bca3e390813ca3d021aa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a2322a3823793c1dd16af892010cd522aeb59d3ccb359e34c603a3ae0181d32\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"message\\\":\\\"rred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:56:52Z is after 2025-08-24T17:21:41Z]\\\\nI0930 13:56:52.715429 6266 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-hmwzh in node crc\\\\nI0930 13:56:52.715417 6266 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"ba175bbe-5cc4-47e6-a32d-57693e1320bd\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-k\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c79f4e921ab2d7e8370fe83552c960147cbbd744f03bca3e390813ca3d021aa3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T13:57:09Z\\\",\\\"message\\\":\\\"irewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 13:57:09.078226 6498 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0930 13:57:09.078250 6498 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0930 13:57:09.078275 6498 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0930 13:57:09.078279 6498 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0930 13:57:09.078296 6498 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0930 13:57:09.078313 6498 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0930 13:57:09.078326 6498 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0930 13:57:09.078344 6498 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0930 13:57:09.078531 6498 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0930 13:57:09.078543 6498 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0930 13:57:09.078563 6498 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0930 13:57:09.078571 6498 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 13:57:09.078579 6498 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0930 13:57:09.078585 6498 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0930 13:57:09.078708 6498 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:09Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.615632 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:09Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.629457 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.629503 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.629517 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.629535 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.629573 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:09Z","lastTransitionTime":"2025-09-30T13:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.631483 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:09Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.644787 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:09Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.657231 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:09Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.672340 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pj2zd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4de996b8-bda8-4142-b5e0-04ed9ae1f327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7286e2d685d6573993e2fa000cb9e08d85c6aa820af507327be52ca50fe58c7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pj2zd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:09Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.690975 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:09Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.707785 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:09Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.731239 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.731276 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.731285 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.731300 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.731310 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:09Z","lastTransitionTime":"2025-09-30T13:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.833757 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.833803 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.833813 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.833829 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.833838 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:09Z","lastTransitionTime":"2025-09-30T13:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.937124 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.937181 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.937198 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.937224 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:09 crc kubenswrapper[4840]: I0930 13:57:09.937242 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:09Z","lastTransitionTime":"2025-09-30T13:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.039227 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.039279 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.039296 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.039319 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.039335 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:10Z","lastTransitionTime":"2025-09-30T13:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.116403 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:57:10 crc kubenswrapper[4840]: E0930 13:57:10.116600 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.142113 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.142171 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.142183 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.142206 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.142218 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:10Z","lastTransitionTime":"2025-09-30T13:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.244974 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.245018 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.245028 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.245042 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.245053 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:10Z","lastTransitionTime":"2025-09-30T13:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.348195 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.348261 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.348282 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.348307 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.348328 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:10Z","lastTransitionTime":"2025-09-30T13:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.450313 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.450348 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.450356 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.450369 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.450378 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:10Z","lastTransitionTime":"2025-09-30T13:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.477329 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kfx69_2ff48c28-d076-46e8-a93f-9630989f81e8/ovnkube-controller/2.log" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.480504 4840 scope.go:117] "RemoveContainer" containerID="c79f4e921ab2d7e8370fe83552c960147cbbd744f03bca3e390813ca3d021aa3" Sep 30 13:57:10 crc kubenswrapper[4840]: E0930 13:57:10.480740 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-kfx69_openshift-ovn-kubernetes(2ff48c28-d076-46e8-a93f-9630989f81e8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.495182 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:10Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.505848 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pj2zd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4de996b8-bda8-4142-b5e0-04ed9ae1f327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7286e2d685d6573993e2fa000cb9e08d85c6aa820af507327be52ca50fe58c7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pj2zd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:10Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.520497 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:10Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.533960 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:10Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.551301 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:10Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.552877 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.552992 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.553073 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.553162 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.553240 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:10Z","lastTransitionTime":"2025-09-30T13:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.569867 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:10Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.585605 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8d388c29baf0ed22d5c9240f2585e4ca133de265227af6c802a169da6650f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:10Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.596632 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b16edc8f720d4296d7fe50fe407f5983ab975818ec6954d1f0e738e5596624a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:10Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.608205 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:10Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.619049 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c2d8050-9531-4413-9d18-14bf3be65acf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34a542c60726f160b423c3e546ff4769a4a7e4a2441b5a09099e4d8c688286c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b5244dfa04a1f1804ef64dc65030d5eaedf679849eec6a9f0731e6cf147f83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7nvs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:10Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.630516 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gp5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1491f559-bc12-4afd-a40c-4eaa40d920a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gp5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:10Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.640681 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49c5addb-0d04-4c3d-b10e-a07d5fec55da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cf852cfa87ef328d172369c26197a9e6a58fe26fa3113469d9275b2e78c30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a08b53dbed5f3a033d2838519733a2a4109ac5a954d9f5fabd0762b2823b090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3f57df6d22c47163751c297aff0ca1dc2bdbf150b9e385e72acdefde92910c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08be626d75cb6bf53342c4ccab7abef9374735cd29d4afb7c7b5b62715db5d22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08be626d75cb6bf53342c4ccab7abef9374735cd29d4afb7c7b5b62715db5d22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:10Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.653219 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f8cd5bd-4a7f-4a5b-ab00-0cc64de9de94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cba320f443ed90a4e755228b663a1da3096e7ea8e068671d1f6273544b6ef60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c507106f39ae2e8acd4268c4e8f28ffd88a0965e69544d8d8a8830e83eceaf9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d62fbefb67a24fedc5f13652596c76f5b288fec58d79b5209f49c79db139d79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61317d26f40a4361a289779bb97b235525b8dddae383f63abfdcab634b633479\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\"\\\\nI0930 13:56:34.911386 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 13:56:34.920195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 13:56:34.920220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 13:56:34.920240 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 13:56:34.920248 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 13:56:34.927953 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 13:56:34.927977 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927984 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 13:56:34.927992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 13:56:34.927996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 13:56:34.928001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 13:56:34.928159 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0930 13:56:34.930945 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759240579\\\\\\\\\\\\\\\" (2025-09-30 13:56:19 +0000 UTC to 2025-10-30 13:56:20 +0000 UTC (now=2025-09-30 13:56:34.930900435 +0000 UTC))\\\\\\\"\\\\nF0930 13:56:34.930995 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1435e2e88d45e5561ff5863095d0f953ef1478d19d87691ab226cad7ae148f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:10Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.654811 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.654840 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.654848 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.654862 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.654872 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:10Z","lastTransitionTime":"2025-09-30T13:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.663798 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:10Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.682811 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84052a424fd6279d70b54902ebdcf430c8aa53b02406b52fc261db9d4d88ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27a94b18137fa991459484406b1c5bd2b2dbb52f527d5f13156ba8e5b8d0146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc4e5d3625aa2d5f94307cedddf72cc48f17f327a044501144e35422d670d555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9549ce85242f1f19b3a52e603caec58bb4c3e2449eb50d8c4525f56dfc93e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a364a13c773fe00ee4e022a286821013d3a681acc16567e66b1c33c7e18ee3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79f4e921ab2d7e8370fe83552c960147cbbd744f03bca3e390813ca3d021aa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c79f4e921ab2d7e8370fe83552c960147cbbd744f03bca3e390813ca3d021aa3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T13:57:09Z\\\",\\\"message\\\":\\\"irewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 13:57:09.078226 6498 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0930 13:57:09.078250 6498 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0930 13:57:09.078275 6498 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0930 13:57:09.078279 6498 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0930 13:57:09.078296 6498 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0930 13:57:09.078313 6498 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0930 13:57:09.078326 6498 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0930 13:57:09.078344 6498 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0930 13:57:09.078531 6498 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0930 13:57:09.078543 6498 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0930 13:57:09.078563 6498 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0930 13:57:09.078571 6498 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 13:57:09.078579 6498 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0930 13:57:09.078585 6498 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0930 13:57:09.078708 6498 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:57:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-kfx69_openshift-ovn-kubernetes(2ff48c28-d076-46e8-a93f-9630989f81e8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:10Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.694878 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:10Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.757039 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.757099 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.757120 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.757152 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.757174 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:10Z","lastTransitionTime":"2025-09-30T13:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.859762 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.859848 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.859873 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.859906 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.859928 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:10Z","lastTransitionTime":"2025-09-30T13:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.912788 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.912878 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.912905 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.912928 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.912947 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:57:10 crc kubenswrapper[4840]: E0930 13:57:10.912975 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:57:42.912952113 +0000 UTC m=+91.542038536 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:57:10 crc kubenswrapper[4840]: E0930 13:57:10.913032 4840 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 13:57:10 crc kubenswrapper[4840]: E0930 13:57:10.913055 4840 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 13:57:10 crc kubenswrapper[4840]: E0930 13:57:10.913065 4840 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 13:57:10 crc kubenswrapper[4840]: E0930 13:57:10.913081 4840 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 13:57:10 crc kubenswrapper[4840]: E0930 13:57:10.913078 4840 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 13:57:10 crc kubenswrapper[4840]: E0930 13:57:10.913082 4840 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 13:57:10 crc kubenswrapper[4840]: E0930 13:57:10.913135 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-30 13:57:42.913118867 +0000 UTC m=+91.542205290 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 13:57:10 crc kubenswrapper[4840]: E0930 13:57:10.913140 4840 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 13:57:10 crc kubenswrapper[4840]: E0930 13:57:10.913064 4840 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 13:57:10 crc kubenswrapper[4840]: E0930 13:57:10.913151 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 13:57:42.913142968 +0000 UTC m=+91.542229391 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 13:57:10 crc kubenswrapper[4840]: E0930 13:57:10.913214 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-30 13:57:42.913202409 +0000 UTC m=+91.542288932 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 13:57:10 crc kubenswrapper[4840]: E0930 13:57:10.913227 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 13:57:42.91322053 +0000 UTC m=+91.542307093 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.962914 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.962956 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.962967 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.962982 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:10 crc kubenswrapper[4840]: I0930 13:57:10.962994 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:10Z","lastTransitionTime":"2025-09-30T13:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.065074 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.065156 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.065174 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.065198 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.065215 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:11Z","lastTransitionTime":"2025-09-30T13:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.116460 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.116517 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.116468 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:57:11 crc kubenswrapper[4840]: E0930 13:57:11.116653 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:57:11 crc kubenswrapper[4840]: E0930 13:57:11.116814 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:57:11 crc kubenswrapper[4840]: E0930 13:57:11.116916 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.167177 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.167246 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.167258 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.167282 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.167295 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:11Z","lastTransitionTime":"2025-09-30T13:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.270761 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.270811 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.270820 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.270835 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.270846 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:11Z","lastTransitionTime":"2025-09-30T13:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.373246 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.373308 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.373318 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.373337 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.373346 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:11Z","lastTransitionTime":"2025-09-30T13:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.476243 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.476314 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.476331 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.476356 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.476376 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:11Z","lastTransitionTime":"2025-09-30T13:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.579423 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.579497 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.579512 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.579531 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.579544 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:11Z","lastTransitionTime":"2025-09-30T13:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.665602 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.682807 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.682893 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.682916 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.682947 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.682970 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:11Z","lastTransitionTime":"2025-09-30T13:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.691123 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:11Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.712844 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:11Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.732001 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49c5addb-0d04-4c3d-b10e-a07d5fec55da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cf852cfa87ef328d172369c26197a9e6a58fe26fa3113469d9275b2e78c30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a08b53dbed5f3a033d2838519733a2a4109ac5a954d9f5fabd0762b2823b090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3f57df6d22c47163751c297aff0ca1dc2bdbf150b9e385e72acdefde92910c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08be626d75cb6bf53342c4ccab7abef9374735cd29d4afb7c7b5b62715db5d22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08be626d75cb6bf53342c4ccab7abef9374735cd29d4afb7c7b5b62715db5d22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:11Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.753231 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f8cd5bd-4a7f-4a5b-ab00-0cc64de9de94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cba320f443ed90a4e755228b663a1da3096e7ea8e068671d1f6273544b6ef60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c507106f39ae2e8acd4268c4e8f28ffd88a0965e69544d8d8a8830e83eceaf9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d62fbefb67a24fedc5f13652596c76f5b288fec58d79b5209f49c79db139d79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61317d26f40a4361a289779bb97b235525b8dddae383f63abfdcab634b633479\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\"\\\\nI0930 13:56:34.911386 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 13:56:34.920195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 13:56:34.920220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 13:56:34.920240 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 13:56:34.920248 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 13:56:34.927953 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 13:56:34.927977 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927984 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 13:56:34.927992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 13:56:34.927996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 13:56:34.928001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 13:56:34.928159 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0930 13:56:34.930945 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759240579\\\\\\\\\\\\\\\" (2025-09-30 13:56:19 +0000 UTC to 2025-10-30 13:56:20 +0000 UTC (now=2025-09-30 13:56:34.930900435 +0000 UTC))\\\\\\\"\\\\nF0930 13:56:34.930995 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1435e2e88d45e5561ff5863095d0f953ef1478d19d87691ab226cad7ae148f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:11Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.766894 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:11Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.780328 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8d388c29baf0ed22d5c9240f2585e4ca133de265227af6c802a169da6650f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:11Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.789184 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.789224 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.789238 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.789254 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.789266 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:11Z","lastTransitionTime":"2025-09-30T13:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.795857 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b16edc8f720d4296d7fe50fe407f5983ab975818ec6954d1f0e738e5596624a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:11Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.805197 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:11Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.816851 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c2d8050-9531-4413-9d18-14bf3be65acf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34a542c60726f160b423c3e546ff4769a4a7e4a2441b5a09099e4d8c688286c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b5244dfa04a1f1804ef64dc65030d5eaedf679849eec6a9f0731e6cf147f83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7nvs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:11Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.832082 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gp5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1491f559-bc12-4afd-a40c-4eaa40d920a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gp5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:11Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.848838 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84052a424fd6279d70b54902ebdcf430c8aa53b02406b52fc261db9d4d88ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27a94b18137fa991459484406b1c5bd2b2dbb52f527d5f13156ba8e5b8d0146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc4e5d3625aa2d5f94307cedddf72cc48f17f327a044501144e35422d670d555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9549ce85242f1f19b3a52e603caec58bb4c3e2449eb50d8c4525f56dfc93e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a364a13c773fe00ee4e022a286821013d3a681acc16567e66b1c33c7e18ee3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79f4e921ab2d7e8370fe83552c960147cbbd744f03bca3e390813ca3d021aa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c79f4e921ab2d7e8370fe83552c960147cbbd744f03bca3e390813ca3d021aa3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T13:57:09Z\\\",\\\"message\\\":\\\"irewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 13:57:09.078226 6498 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0930 13:57:09.078250 6498 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0930 13:57:09.078275 6498 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0930 13:57:09.078279 6498 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0930 13:57:09.078296 6498 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0930 13:57:09.078313 6498 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0930 13:57:09.078326 6498 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0930 13:57:09.078344 6498 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0930 13:57:09.078531 6498 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0930 13:57:09.078543 6498 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0930 13:57:09.078563 6498 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0930 13:57:09.078571 6498 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 13:57:09.078579 6498 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0930 13:57:09.078585 6498 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0930 13:57:09.078708 6498 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:57:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-kfx69_openshift-ovn-kubernetes(2ff48c28-d076-46e8-a93f-9630989f81e8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:11Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.863747 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:11Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.878119 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:11Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.891665 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:11Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.892238 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.892272 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.892283 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.892300 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.892311 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:11Z","lastTransitionTime":"2025-09-30T13:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.907165 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:11Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.920997 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pj2zd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4de996b8-bda8-4142-b5e0-04ed9ae1f327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7286e2d685d6573993e2fa000cb9e08d85c6aa820af507327be52ca50fe58c7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pj2zd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:11Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.994898 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.994948 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.994957 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.994971 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:11 crc kubenswrapper[4840]: I0930 13:57:11.994982 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:11Z","lastTransitionTime":"2025-09-30T13:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.097775 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.097822 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.097832 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.097849 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.097860 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:12Z","lastTransitionTime":"2025-09-30T13:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.115528 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:57:12 crc kubenswrapper[4840]: E0930 13:57:12.115813 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.128541 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:12Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.141212 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:12Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.153700 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49c5addb-0d04-4c3d-b10e-a07d5fec55da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cf852cfa87ef328d172369c26197a9e6a58fe26fa3113469d9275b2e78c30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a08b53dbed5f3a033d2838519733a2a4109ac5a954d9f5fabd0762b2823b090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3f57df6d22c47163751c297aff0ca1dc2bdbf150b9e385e72acdefde92910c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08be626d75cb6bf53342c4ccab7abef9374735cd29d4afb7c7b5b62715db5d22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08be626d75cb6bf53342c4ccab7abef9374735cd29d4afb7c7b5b62715db5d22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:12Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.167206 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f8cd5bd-4a7f-4a5b-ab00-0cc64de9de94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cba320f443ed90a4e755228b663a1da3096e7ea8e068671d1f6273544b6ef60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c507106f39ae2e8acd4268c4e8f28ffd88a0965e69544d8d8a8830e83eceaf9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d62fbefb67a24fedc5f13652596c76f5b288fec58d79b5209f49c79db139d79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61317d26f40a4361a289779bb97b235525b8dddae383f63abfdcab634b633479\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\"\\\\nI0930 13:56:34.911386 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 13:56:34.920195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 13:56:34.920220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 13:56:34.920240 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 13:56:34.920248 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 13:56:34.927953 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 13:56:34.927977 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927984 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 13:56:34.927992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 13:56:34.927996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 13:56:34.928001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 13:56:34.928159 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0930 13:56:34.930945 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759240579\\\\\\\\\\\\\\\" (2025-09-30 13:56:19 +0000 UTC to 2025-10-30 13:56:20 +0000 UTC (now=2025-09-30 13:56:34.930900435 +0000 UTC))\\\\\\\"\\\\nF0930 13:56:34.930995 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1435e2e88d45e5561ff5863095d0f953ef1478d19d87691ab226cad7ae148f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:12Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.179212 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:12Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.197405 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8d388c29baf0ed22d5c9240f2585e4ca133de265227af6c802a169da6650f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:12Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.199953 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.200003 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.200021 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.200045 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.200060 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:12Z","lastTransitionTime":"2025-09-30T13:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.210493 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b16edc8f720d4296d7fe50fe407f5983ab975818ec6954d1f0e738e5596624a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:12Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.221621 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:12Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.237035 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c2d8050-9531-4413-9d18-14bf3be65acf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34a542c60726f160b423c3e546ff4769a4a7e4a2441b5a09099e4d8c688286c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b5244dfa04a1f1804ef64dc65030d5eaedf679849eec6a9f0731e6cf147f83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7nvs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:12Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.255994 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gp5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1491f559-bc12-4afd-a40c-4eaa40d920a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gp5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:12Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.272648 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84052a424fd6279d70b54902ebdcf430c8aa53b02406b52fc261db9d4d88ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27a94b18137fa991459484406b1c5bd2b2dbb52f527d5f13156ba8e5b8d0146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc4e5d3625aa2d5f94307cedddf72cc48f17f327a044501144e35422d670d555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9549ce85242f1f19b3a52e603caec58bb4c3e2449eb50d8c4525f56dfc93e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a364a13c773fe00ee4e022a286821013d3a681acc16567e66b1c33c7e18ee3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79f4e921ab2d7e8370fe83552c960147cbbd744f03bca3e390813ca3d021aa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c79f4e921ab2d7e8370fe83552c960147cbbd744f03bca3e390813ca3d021aa3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T13:57:09Z\\\",\\\"message\\\":\\\"irewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 13:57:09.078226 6498 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0930 13:57:09.078250 6498 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0930 13:57:09.078275 6498 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0930 13:57:09.078279 6498 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0930 13:57:09.078296 6498 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0930 13:57:09.078313 6498 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0930 13:57:09.078326 6498 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0930 13:57:09.078344 6498 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0930 13:57:09.078531 6498 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0930 13:57:09.078543 6498 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0930 13:57:09.078563 6498 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0930 13:57:09.078571 6498 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 13:57:09.078579 6498 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0930 13:57:09.078585 6498 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0930 13:57:09.078708 6498 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:57:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-kfx69_openshift-ovn-kubernetes(2ff48c28-d076-46e8-a93f-9630989f81e8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:12Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.283620 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:12Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.294059 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:12Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.302233 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.302264 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.302274 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.302287 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.302297 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:12Z","lastTransitionTime":"2025-09-30T13:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.306066 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:12Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.315539 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:12Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.322680 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pj2zd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4de996b8-bda8-4142-b5e0-04ed9ae1f327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7286e2d685d6573993e2fa000cb9e08d85c6aa820af507327be52ca50fe58c7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pj2zd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:12Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.405127 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.405192 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.405209 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.405239 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.405258 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:12Z","lastTransitionTime":"2025-09-30T13:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.509101 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.509176 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.509204 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.509238 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.509299 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:12Z","lastTransitionTime":"2025-09-30T13:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.612672 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.612759 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.612787 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.612820 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.612846 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:12Z","lastTransitionTime":"2025-09-30T13:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.716644 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.717130 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.717146 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.717173 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.717189 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:12Z","lastTransitionTime":"2025-09-30T13:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.820209 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.820253 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.820261 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.820275 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.820284 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:12Z","lastTransitionTime":"2025-09-30T13:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.923282 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.923333 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.923355 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.923374 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:12 crc kubenswrapper[4840]: I0930 13:57:12.923389 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:12Z","lastTransitionTime":"2025-09-30T13:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.026721 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.026764 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.026773 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.026788 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.026797 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:13Z","lastTransitionTime":"2025-09-30T13:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.116218 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.116260 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.116276 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:57:13 crc kubenswrapper[4840]: E0930 13:57:13.116427 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:57:13 crc kubenswrapper[4840]: E0930 13:57:13.116585 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:57:13 crc kubenswrapper[4840]: E0930 13:57:13.116759 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.130355 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.130405 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.130416 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.130433 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.130442 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:13Z","lastTransitionTime":"2025-09-30T13:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.233117 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.233173 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.233184 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.233201 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.233215 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:13Z","lastTransitionTime":"2025-09-30T13:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.335524 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.335597 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.335610 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.335630 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.335640 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:13Z","lastTransitionTime":"2025-09-30T13:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.439324 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.439406 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.439420 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.439446 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.439460 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:13Z","lastTransitionTime":"2025-09-30T13:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.542106 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.542167 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.542181 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.542202 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.542215 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:13Z","lastTransitionTime":"2025-09-30T13:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.645666 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.645712 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.645724 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.645743 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.645756 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:13Z","lastTransitionTime":"2025-09-30T13:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.747776 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.747817 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.747829 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.747846 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.747859 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:13Z","lastTransitionTime":"2025-09-30T13:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.850460 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.850503 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.850517 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.850546 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.850607 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:13Z","lastTransitionTime":"2025-09-30T13:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.953374 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.953417 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.953425 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.953437 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:13 crc kubenswrapper[4840]: I0930 13:57:13.953447 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:13Z","lastTransitionTime":"2025-09-30T13:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.056255 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.056304 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.056313 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.056331 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.056340 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:14Z","lastTransitionTime":"2025-09-30T13:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.115924 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:57:14 crc kubenswrapper[4840]: E0930 13:57:14.116100 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.159955 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.160017 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.160036 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.160060 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.160077 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:14Z","lastTransitionTime":"2025-09-30T13:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.263069 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.263150 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.263168 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.263192 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.263217 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:14Z","lastTransitionTime":"2025-09-30T13:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.366807 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.366876 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.366900 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.366929 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.366952 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:14Z","lastTransitionTime":"2025-09-30T13:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.470490 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.470606 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.470632 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.470676 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.470693 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:14Z","lastTransitionTime":"2025-09-30T13:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.573478 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.573532 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.573541 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.573577 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.573596 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:14Z","lastTransitionTime":"2025-09-30T13:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.675823 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.675870 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.675884 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.675901 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.675913 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:14Z","lastTransitionTime":"2025-09-30T13:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.778709 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.778748 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.778759 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.778774 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.778784 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:14Z","lastTransitionTime":"2025-09-30T13:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.881494 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.881581 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.881604 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.881625 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.881643 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:14Z","lastTransitionTime":"2025-09-30T13:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.984213 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.984265 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.984284 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.984319 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:14 crc kubenswrapper[4840]: I0930 13:57:14.984352 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:14Z","lastTransitionTime":"2025-09-30T13:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.086473 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.086525 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.086536 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.086574 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.086589 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:15Z","lastTransitionTime":"2025-09-30T13:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.116290 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.116341 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.116405 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:57:15 crc kubenswrapper[4840]: E0930 13:57:15.116483 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:57:15 crc kubenswrapper[4840]: E0930 13:57:15.116635 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:57:15 crc kubenswrapper[4840]: E0930 13:57:15.116706 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.188964 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.189002 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.189012 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.189028 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.189040 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:15Z","lastTransitionTime":"2025-09-30T13:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.291063 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.291099 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.291111 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.291126 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.291137 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:15Z","lastTransitionTime":"2025-09-30T13:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.396464 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.396610 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.396643 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.396678 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.396712 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:15Z","lastTransitionTime":"2025-09-30T13:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.499309 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.499352 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.499363 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.499381 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.499392 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:15Z","lastTransitionTime":"2025-09-30T13:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.601206 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.601245 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.601254 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.601268 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.601279 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:15Z","lastTransitionTime":"2025-09-30T13:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.703419 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.703473 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.703484 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.703511 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.703525 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:15Z","lastTransitionTime":"2025-09-30T13:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.806597 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.806684 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.806719 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.806753 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.806778 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:15Z","lastTransitionTime":"2025-09-30T13:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.909662 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.909700 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.909710 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.909726 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:15 crc kubenswrapper[4840]: I0930 13:57:15.909739 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:15Z","lastTransitionTime":"2025-09-30T13:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.011723 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.011775 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.011786 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.011806 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.011816 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:16Z","lastTransitionTime":"2025-09-30T13:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.114204 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.114246 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.114258 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.114274 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.114286 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:16Z","lastTransitionTime":"2025-09-30T13:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.115435 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:57:16 crc kubenswrapper[4840]: E0930 13:57:16.115631 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.170972 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.171006 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.171015 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.171030 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.171039 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:16Z","lastTransitionTime":"2025-09-30T13:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:16 crc kubenswrapper[4840]: E0930 13:57:16.183249 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5\\\",\\\"systemUUID\\\":\\\"f41e0351-4eda-411a-9a99-43a1952f3d34\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:16Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.186713 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.186745 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.186757 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.186773 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.186782 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:16Z","lastTransitionTime":"2025-09-30T13:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:16 crc kubenswrapper[4840]: E0930 13:57:16.198949 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5\\\",\\\"systemUUID\\\":\\\"f41e0351-4eda-411a-9a99-43a1952f3d34\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:16Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.202059 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.202106 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.202121 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.202145 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.202161 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:16Z","lastTransitionTime":"2025-09-30T13:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:16 crc kubenswrapper[4840]: E0930 13:57:16.233172 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5\\\",\\\"systemUUID\\\":\\\"f41e0351-4eda-411a-9a99-43a1952f3d34\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:16Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.239001 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.239031 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.239039 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.239053 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.239062 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:16Z","lastTransitionTime":"2025-09-30T13:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:16 crc kubenswrapper[4840]: E0930 13:57:16.255430 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5\\\",\\\"systemUUID\\\":\\\"f41e0351-4eda-411a-9a99-43a1952f3d34\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:16Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.258793 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.258826 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.258835 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.258848 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.258857 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:16Z","lastTransitionTime":"2025-09-30T13:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:16 crc kubenswrapper[4840]: E0930 13:57:16.272937 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5\\\",\\\"systemUUID\\\":\\\"f41e0351-4eda-411a-9a99-43a1952f3d34\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:16Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:16 crc kubenswrapper[4840]: E0930 13:57:16.273090 4840 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.274583 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.274612 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.274622 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.274637 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.274646 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:16Z","lastTransitionTime":"2025-09-30T13:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.376799 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.376860 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.376873 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.376889 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.376898 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:16Z","lastTransitionTime":"2025-09-30T13:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.478833 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.478913 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.478931 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.478958 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.478988 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:16Z","lastTransitionTime":"2025-09-30T13:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.581453 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.581494 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.581519 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.581543 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.581583 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:16Z","lastTransitionTime":"2025-09-30T13:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.683625 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.683720 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.683733 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.683748 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.683760 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:16Z","lastTransitionTime":"2025-09-30T13:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.786107 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.786148 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.786158 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.786174 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.786184 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:16Z","lastTransitionTime":"2025-09-30T13:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.888691 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.888726 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.888736 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.888751 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.888762 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:16Z","lastTransitionTime":"2025-09-30T13:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.990511 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.990600 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.990612 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.990630 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:16 crc kubenswrapper[4840]: I0930 13:57:16.990642 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:16Z","lastTransitionTime":"2025-09-30T13:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.093454 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.093499 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.093509 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.093526 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.093537 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:17Z","lastTransitionTime":"2025-09-30T13:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.116048 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.116081 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.116118 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:57:17 crc kubenswrapper[4840]: E0930 13:57:17.116270 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:57:17 crc kubenswrapper[4840]: E0930 13:57:17.116427 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:57:17 crc kubenswrapper[4840]: E0930 13:57:17.116461 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.195692 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.195752 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.195768 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.195790 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.195804 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:17Z","lastTransitionTime":"2025-09-30T13:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.298188 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.298217 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.298225 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.298239 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.298247 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:17Z","lastTransitionTime":"2025-09-30T13:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.401809 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.402374 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.402387 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.402403 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.402414 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:17Z","lastTransitionTime":"2025-09-30T13:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.504252 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.504327 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.504348 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.504377 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.504394 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:17Z","lastTransitionTime":"2025-09-30T13:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.608105 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.608173 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.608190 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.608214 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.608232 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:17Z","lastTransitionTime":"2025-09-30T13:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.710351 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.710405 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.710415 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.710428 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.710436 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:17Z","lastTransitionTime":"2025-09-30T13:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.812988 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.813042 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.813053 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.813072 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.813083 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:17Z","lastTransitionTime":"2025-09-30T13:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.917013 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.917091 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.917107 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.917130 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:17 crc kubenswrapper[4840]: I0930 13:57:17.917146 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:17Z","lastTransitionTime":"2025-09-30T13:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.020142 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.020177 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.020186 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.020212 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.020223 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:18Z","lastTransitionTime":"2025-09-30T13:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.116023 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:57:18 crc kubenswrapper[4840]: E0930 13:57:18.116197 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.121689 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.121760 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.121777 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.121803 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.121821 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:18Z","lastTransitionTime":"2025-09-30T13:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.225813 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.225859 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.225874 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.225895 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.225908 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:18Z","lastTransitionTime":"2025-09-30T13:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.328391 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.328431 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.328441 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.328456 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.328466 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:18Z","lastTransitionTime":"2025-09-30T13:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.430481 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.430529 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.430542 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.430589 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.430602 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:18Z","lastTransitionTime":"2025-09-30T13:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.533194 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.533237 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.533247 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.533261 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.533272 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:18Z","lastTransitionTime":"2025-09-30T13:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.635769 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.635824 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.635843 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.635866 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.635884 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:18Z","lastTransitionTime":"2025-09-30T13:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.739229 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.739269 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.739285 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.739302 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.739311 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:18Z","lastTransitionTime":"2025-09-30T13:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.842057 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.842125 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.842141 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.842158 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.842168 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:18Z","lastTransitionTime":"2025-09-30T13:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.945757 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.945805 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.945817 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.945836 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:18 crc kubenswrapper[4840]: I0930 13:57:18.945847 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:18Z","lastTransitionTime":"2025-09-30T13:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.048876 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.048950 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.048964 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.048990 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.049006 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:19Z","lastTransitionTime":"2025-09-30T13:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.115915 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.115926 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:57:19 crc kubenswrapper[4840]: E0930 13:57:19.116134 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.115951 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:57:19 crc kubenswrapper[4840]: E0930 13:57:19.116264 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:57:19 crc kubenswrapper[4840]: E0930 13:57:19.116340 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.151029 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.151081 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.151092 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.151109 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.151122 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:19Z","lastTransitionTime":"2025-09-30T13:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.253540 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.253603 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.253617 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.253637 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.253651 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:19Z","lastTransitionTime":"2025-09-30T13:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.356068 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.356111 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.356124 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.356140 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.356149 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:19Z","lastTransitionTime":"2025-09-30T13:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.459001 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.459056 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.459066 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.459082 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.459093 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:19Z","lastTransitionTime":"2025-09-30T13:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.561098 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.561147 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.561159 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.561183 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.561198 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:19Z","lastTransitionTime":"2025-09-30T13:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.664075 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.664116 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.664127 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.664144 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.664154 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:19Z","lastTransitionTime":"2025-09-30T13:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.766890 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.766939 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.766951 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.766970 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.766981 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:19Z","lastTransitionTime":"2025-09-30T13:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.869458 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.869499 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.869509 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.869524 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.869535 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:19Z","lastTransitionTime":"2025-09-30T13:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.971940 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.971979 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.971991 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.972008 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:19 crc kubenswrapper[4840]: I0930 13:57:19.972019 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:19Z","lastTransitionTime":"2025-09-30T13:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.075204 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.075245 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.075253 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.075268 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.075278 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:20Z","lastTransitionTime":"2025-09-30T13:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.115927 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:57:20 crc kubenswrapper[4840]: E0930 13:57:20.116068 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.177739 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.177784 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.177796 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.177815 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.177829 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:20Z","lastTransitionTime":"2025-09-30T13:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.280225 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.280257 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.280268 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.280281 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.280292 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:20Z","lastTransitionTime":"2025-09-30T13:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.383450 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.383483 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.383493 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.383509 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.383519 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:20Z","lastTransitionTime":"2025-09-30T13:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.486487 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.486545 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.486603 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.486643 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.486668 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:20Z","lastTransitionTime":"2025-09-30T13:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.589453 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.589481 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.589489 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.589504 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.589513 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:20Z","lastTransitionTime":"2025-09-30T13:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.691769 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.691815 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.691825 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.691839 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.691849 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:20Z","lastTransitionTime":"2025-09-30T13:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.793963 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.794011 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.794027 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.794049 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.794063 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:20Z","lastTransitionTime":"2025-09-30T13:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.896602 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.896647 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.896658 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.896673 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.896682 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:20Z","lastTransitionTime":"2025-09-30T13:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.998756 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.998834 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.998858 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.998890 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:20 crc kubenswrapper[4840]: I0930 13:57:20.998912 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:20Z","lastTransitionTime":"2025-09-30T13:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.101363 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.101452 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.101463 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.101482 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.101498 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:21Z","lastTransitionTime":"2025-09-30T13:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.115613 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.115766 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:57:21 crc kubenswrapper[4840]: E0930 13:57:21.115854 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.115899 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:57:21 crc kubenswrapper[4840]: E0930 13:57:21.115948 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:57:21 crc kubenswrapper[4840]: E0930 13:57:21.116001 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.204044 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.204083 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.204091 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.204104 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.204112 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:21Z","lastTransitionTime":"2025-09-30T13:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.306064 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.306096 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.306104 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.306117 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.306126 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:21Z","lastTransitionTime":"2025-09-30T13:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.408922 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.408961 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.408970 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.408985 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.408996 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:21Z","lastTransitionTime":"2025-09-30T13:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.511518 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.511568 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.511579 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.511596 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.511607 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:21Z","lastTransitionTime":"2025-09-30T13:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.614850 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.614899 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.614910 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.614930 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.614942 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:21Z","lastTransitionTime":"2025-09-30T13:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.717923 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.717999 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.718025 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.718046 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.718096 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:21Z","lastTransitionTime":"2025-09-30T13:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.754117 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1491f559-bc12-4afd-a40c-4eaa40d920a8-metrics-certs\") pod \"network-metrics-daemon-4gp5x\" (UID: \"1491f559-bc12-4afd-a40c-4eaa40d920a8\") " pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:57:21 crc kubenswrapper[4840]: E0930 13:57:21.754388 4840 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 13:57:21 crc kubenswrapper[4840]: E0930 13:57:21.754529 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1491f559-bc12-4afd-a40c-4eaa40d920a8-metrics-certs podName:1491f559-bc12-4afd-a40c-4eaa40d920a8 nodeName:}" failed. No retries permitted until 2025-09-30 13:57:53.75449437 +0000 UTC m=+102.383580983 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1491f559-bc12-4afd-a40c-4eaa40d920a8-metrics-certs") pod "network-metrics-daemon-4gp5x" (UID: "1491f559-bc12-4afd-a40c-4eaa40d920a8") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.820920 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.820953 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.820963 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.820983 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.821002 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:21Z","lastTransitionTime":"2025-09-30T13:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.923801 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.923851 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.923863 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.923879 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:21 crc kubenswrapper[4840]: I0930 13:57:21.923891 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:21Z","lastTransitionTime":"2025-09-30T13:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.030266 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.030316 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.030328 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.030348 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.030363 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:22Z","lastTransitionTime":"2025-09-30T13:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.115965 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:57:22 crc kubenswrapper[4840]: E0930 13:57:22.116129 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.132338 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:22Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.133030 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.133055 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.133067 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.133083 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.133095 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:22Z","lastTransitionTime":"2025-09-30T13:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.145501 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:22Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.159177 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f8cd5bd-4a7f-4a5b-ab00-0cc64de9de94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cba320f443ed90a4e755228b663a1da3096e7ea8e068671d1f6273544b6ef60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c507106f39ae2e8acd4268c4e8f28ffd88a0965e69544d8d8a8830e83eceaf9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d62fbefb67a24fedc5f13652596c76f5b288fec58d79b5209f49c79db139d79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61317d26f40a4361a289779bb97b235525b8dddae383f63abfdcab634b633479\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\"\\\\nI0930 13:56:34.911386 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 13:56:34.920195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 13:56:34.920220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 13:56:34.920240 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 13:56:34.920248 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 13:56:34.927953 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 13:56:34.927977 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927984 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 13:56:34.927992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 13:56:34.927996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 13:56:34.928001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 13:56:34.928159 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0930 13:56:34.930945 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759240579\\\\\\\\\\\\\\\" (2025-09-30 13:56:19 +0000 UTC to 2025-10-30 13:56:20 +0000 UTC (now=2025-09-30 13:56:34.930900435 +0000 UTC))\\\\\\\"\\\\nF0930 13:56:34.930995 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1435e2e88d45e5561ff5863095d0f953ef1478d19d87691ab226cad7ae148f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:22Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.169994 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:22Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.183055 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8d388c29baf0ed22d5c9240f2585e4ca133de265227af6c802a169da6650f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:22Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.195536 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b16edc8f720d4296d7fe50fe407f5983ab975818ec6954d1f0e738e5596624a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:22Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.205936 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:22Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.217594 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c2d8050-9531-4413-9d18-14bf3be65acf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34a542c60726f160b423c3e546ff4769a4a7e4a2441b5a09099e4d8c688286c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b5244dfa04a1f1804ef64dc65030d5eaedf679849eec6a9f0731e6cf147f83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7nvs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:22Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.228355 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gp5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1491f559-bc12-4afd-a40c-4eaa40d920a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gp5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:22Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.236373 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.236417 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.236428 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.236449 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.236463 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:22Z","lastTransitionTime":"2025-09-30T13:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.241310 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49c5addb-0d04-4c3d-b10e-a07d5fec55da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cf852cfa87ef328d172369c26197a9e6a58fe26fa3113469d9275b2e78c30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a08b53dbed5f3a033d2838519733a2a4109ac5a954d9f5fabd0762b2823b090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3f57df6d22c47163751c297aff0ca1dc2bdbf150b9e385e72acdefde92910c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08be626d75cb6bf53342c4ccab7abef9374735cd29d4afb7c7b5b62715db5d22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08be626d75cb6bf53342c4ccab7abef9374735cd29d4afb7c7b5b62715db5d22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:22Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.258293 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84052a424fd6279d70b54902ebdcf430c8aa53b02406b52fc261db9d4d88ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27a94b18137fa991459484406b1c5bd2b2dbb52f527d5f13156ba8e5b8d0146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc4e5d3625aa2d5f94307cedddf72cc48f17f327a044501144e35422d670d555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9549ce85242f1f19b3a52e603caec58bb4c3e2449eb50d8c4525f56dfc93e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a364a13c773fe00ee4e022a286821013d3a681acc16567e66b1c33c7e18ee3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79f4e921ab2d7e8370fe83552c960147cbbd744f03bca3e390813ca3d021aa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c79f4e921ab2d7e8370fe83552c960147cbbd744f03bca3e390813ca3d021aa3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T13:57:09Z\\\",\\\"message\\\":\\\"irewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 13:57:09.078226 6498 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0930 13:57:09.078250 6498 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0930 13:57:09.078275 6498 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0930 13:57:09.078279 6498 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0930 13:57:09.078296 6498 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0930 13:57:09.078313 6498 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0930 13:57:09.078326 6498 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0930 13:57:09.078344 6498 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0930 13:57:09.078531 6498 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0930 13:57:09.078543 6498 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0930 13:57:09.078563 6498 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0930 13:57:09.078571 6498 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 13:57:09.078579 6498 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0930 13:57:09.078585 6498 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0930 13:57:09.078708 6498 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:57:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-kfx69_openshift-ovn-kubernetes(2ff48c28-d076-46e8-a93f-9630989f81e8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:22Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.277157 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:22Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.289440 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:22Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.300204 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:22Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.312490 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:22Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.323333 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pj2zd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4de996b8-bda8-4142-b5e0-04ed9ae1f327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7286e2d685d6573993e2fa000cb9e08d85c6aa820af507327be52ca50fe58c7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pj2zd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:22Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.339191 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.339255 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.339268 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.339288 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.339301 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:22Z","lastTransitionTime":"2025-09-30T13:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.441758 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.441966 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.442072 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.442239 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.442418 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:22Z","lastTransitionTime":"2025-09-30T13:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.545420 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.545469 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.545485 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.545503 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.545516 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:22Z","lastTransitionTime":"2025-09-30T13:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.648653 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.648700 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.648716 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.648743 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.648756 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:22Z","lastTransitionTime":"2025-09-30T13:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.751132 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.751175 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.751184 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.751200 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.751210 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:22Z","lastTransitionTime":"2025-09-30T13:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.853718 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.853783 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.853800 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.853831 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.853848 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:22Z","lastTransitionTime":"2025-09-30T13:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.957042 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.957128 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.957153 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.957200 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:22 crc kubenswrapper[4840]: I0930 13:57:22.957227 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:22Z","lastTransitionTime":"2025-09-30T13:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.059924 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.059983 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.060001 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.060028 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.060049 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:23Z","lastTransitionTime":"2025-09-30T13:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.115871 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.115972 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.115870 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:57:23 crc kubenswrapper[4840]: E0930 13:57:23.116075 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:57:23 crc kubenswrapper[4840]: E0930 13:57:23.116168 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:57:23 crc kubenswrapper[4840]: E0930 13:57:23.116365 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.163363 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.163397 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.163405 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.163418 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.163426 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:23Z","lastTransitionTime":"2025-09-30T13:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.266497 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.266605 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.266630 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.266661 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.266682 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:23Z","lastTransitionTime":"2025-09-30T13:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.369773 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.369824 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.369839 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.369859 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.369877 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:23Z","lastTransitionTime":"2025-09-30T13:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.473234 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.473260 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.473268 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.473280 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.473305 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:23Z","lastTransitionTime":"2025-09-30T13:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.575455 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.575505 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.575519 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.575540 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.575589 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:23Z","lastTransitionTime":"2025-09-30T13:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.677539 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.677595 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.677606 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.677621 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.677672 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:23Z","lastTransitionTime":"2025-09-30T13:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.780317 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.780354 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.780363 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.780379 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.780388 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:23Z","lastTransitionTime":"2025-09-30T13:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.883406 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.883442 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.883468 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.883481 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.883489 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:23Z","lastTransitionTime":"2025-09-30T13:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.986201 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.986231 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.986239 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.986253 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:23 crc kubenswrapper[4840]: I0930 13:57:23.986264 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:23Z","lastTransitionTime":"2025-09-30T13:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.088310 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.088355 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.088370 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.088393 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.088406 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:24Z","lastTransitionTime":"2025-09-30T13:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.115943 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:57:24 crc kubenswrapper[4840]: E0930 13:57:24.116360 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.116653 4840 scope.go:117] "RemoveContainer" containerID="c79f4e921ab2d7e8370fe83552c960147cbbd744f03bca3e390813ca3d021aa3" Sep 30 13:57:24 crc kubenswrapper[4840]: E0930 13:57:24.116928 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-kfx69_openshift-ovn-kubernetes(2ff48c28-d076-46e8-a93f-9630989f81e8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.190708 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.190760 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.190768 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.190783 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.190792 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:24Z","lastTransitionTime":"2025-09-30T13:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.293710 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.293741 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.293749 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.293764 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.293773 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:24Z","lastTransitionTime":"2025-09-30T13:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.396391 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.396725 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.396814 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.396916 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.397016 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:24Z","lastTransitionTime":"2025-09-30T13:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.499126 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.499363 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.499487 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.499606 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.499688 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:24Z","lastTransitionTime":"2025-09-30T13:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.602633 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.602733 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.602752 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.602780 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.602801 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:24Z","lastTransitionTime":"2025-09-30T13:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.705074 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.705143 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.705157 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.705183 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.705197 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:24Z","lastTransitionTime":"2025-09-30T13:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.806910 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.806961 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.806976 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.806998 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.807013 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:24Z","lastTransitionTime":"2025-09-30T13:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.910297 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.910336 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.910356 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.910373 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:24 crc kubenswrapper[4840]: I0930 13:57:24.910384 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:24Z","lastTransitionTime":"2025-09-30T13:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.013839 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.013864 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.013871 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.013883 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.013891 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:25Z","lastTransitionTime":"2025-09-30T13:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.115404 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:57:25 crc kubenswrapper[4840]: E0930 13:57:25.115529 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.115735 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:57:25 crc kubenswrapper[4840]: E0930 13:57:25.115797 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.115914 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:57:25 crc kubenswrapper[4840]: E0930 13:57:25.115978 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.116407 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.116435 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.116446 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.116459 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.116470 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:25Z","lastTransitionTime":"2025-09-30T13:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.218754 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.218792 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.218806 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.218824 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.218835 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:25Z","lastTransitionTime":"2025-09-30T13:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.321541 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.321599 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.321610 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.321624 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.321634 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:25Z","lastTransitionTime":"2025-09-30T13:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.424660 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.424715 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.424726 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.424743 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.424753 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:25Z","lastTransitionTime":"2025-09-30T13:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.526867 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.526952 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.526974 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.527005 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.527028 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:25Z","lastTransitionTime":"2025-09-30T13:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.629902 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.629948 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.629962 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.629978 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.629990 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:25Z","lastTransitionTime":"2025-09-30T13:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.732251 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.732317 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.732340 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.732370 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.732393 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:25Z","lastTransitionTime":"2025-09-30T13:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.835195 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.835500 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.835643 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.835738 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.835832 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:25Z","lastTransitionTime":"2025-09-30T13:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.939410 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.939694 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.939769 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.939837 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:25 crc kubenswrapper[4840]: I0930 13:57:25.939934 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:25Z","lastTransitionTime":"2025-09-30T13:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.042366 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.042405 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.042415 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.042431 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.042444 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:26Z","lastTransitionTime":"2025-09-30T13:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.115483 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:57:26 crc kubenswrapper[4840]: E0930 13:57:26.115948 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.144821 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.144859 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.144875 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.144889 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.144899 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:26Z","lastTransitionTime":"2025-09-30T13:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.247405 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.247452 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.247461 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.247477 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.247496 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:26Z","lastTransitionTime":"2025-09-30T13:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.350772 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.350823 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.350833 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.350849 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.350861 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:26Z","lastTransitionTime":"2025-09-30T13:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.454183 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.454240 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.454249 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.454264 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.454275 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:26Z","lastTransitionTime":"2025-09-30T13:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.557139 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.557267 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.557289 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.557319 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.557341 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:26Z","lastTransitionTime":"2025-09-30T13:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.630179 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.630231 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.630242 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.630260 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.630278 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:26Z","lastTransitionTime":"2025-09-30T13:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:26 crc kubenswrapper[4840]: E0930 13:57:26.651131 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5\\\",\\\"systemUUID\\\":\\\"f41e0351-4eda-411a-9a99-43a1952f3d34\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:26Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.656941 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.657035 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.657055 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.657081 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.657101 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:26Z","lastTransitionTime":"2025-09-30T13:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:26 crc kubenswrapper[4840]: E0930 13:57:26.678695 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5\\\",\\\"systemUUID\\\":\\\"f41e0351-4eda-411a-9a99-43a1952f3d34\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:26Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.683838 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.683920 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.683935 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.683958 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.683975 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:26Z","lastTransitionTime":"2025-09-30T13:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:26 crc kubenswrapper[4840]: E0930 13:57:26.701777 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5\\\",\\\"systemUUID\\\":\\\"f41e0351-4eda-411a-9a99-43a1952f3d34\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:26Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.706712 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.706774 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.706816 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.706848 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.706861 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:26Z","lastTransitionTime":"2025-09-30T13:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:26 crc kubenswrapper[4840]: E0930 13:57:26.721234 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5\\\",\\\"systemUUID\\\":\\\"f41e0351-4eda-411a-9a99-43a1952f3d34\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:26Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.725080 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.725163 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.725186 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.725221 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.725243 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:26Z","lastTransitionTime":"2025-09-30T13:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:26 crc kubenswrapper[4840]: E0930 13:57:26.740722 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5\\\",\\\"systemUUID\\\":\\\"f41e0351-4eda-411a-9a99-43a1952f3d34\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:26Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:26 crc kubenswrapper[4840]: E0930 13:57:26.740886 4840 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.742878 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.742916 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.742928 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.742941 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.742953 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:26Z","lastTransitionTime":"2025-09-30T13:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.845304 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.845460 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.845473 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.845491 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.845504 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:26Z","lastTransitionTime":"2025-09-30T13:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.948790 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.948855 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.948866 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.948885 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:26 crc kubenswrapper[4840]: I0930 13:57:26.948897 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:26Z","lastTransitionTime":"2025-09-30T13:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.052035 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.052101 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.052115 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.052140 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.052158 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:27Z","lastTransitionTime":"2025-09-30T13:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.115948 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.116016 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.116049 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:57:27 crc kubenswrapper[4840]: E0930 13:57:27.116191 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:57:27 crc kubenswrapper[4840]: E0930 13:57:27.116338 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:57:27 crc kubenswrapper[4840]: E0930 13:57:27.116638 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.155374 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.155439 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.155453 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.155475 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.155492 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:27Z","lastTransitionTime":"2025-09-30T13:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.258689 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.258743 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.258757 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.258783 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.258798 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:27Z","lastTransitionTime":"2025-09-30T13:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.361129 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.361205 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.361222 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.361247 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.361267 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:27Z","lastTransitionTime":"2025-09-30T13:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.464609 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.464679 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.464697 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.464724 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.464740 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:27Z","lastTransitionTime":"2025-09-30T13:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.567756 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.567807 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.567819 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.567836 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.567847 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:27Z","lastTransitionTime":"2025-09-30T13:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.669973 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.670016 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.670027 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.670042 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.670054 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:27Z","lastTransitionTime":"2025-09-30T13:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.773105 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.773161 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.773176 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.773199 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.773214 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:27Z","lastTransitionTime":"2025-09-30T13:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.875516 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.875685 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.875700 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.875715 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.875724 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:27Z","lastTransitionTime":"2025-09-30T13:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.978210 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.978248 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.978256 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.978269 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:27 crc kubenswrapper[4840]: I0930 13:57:27.978278 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:27Z","lastTransitionTime":"2025-09-30T13:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.080465 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.080514 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.080526 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.080542 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.080570 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:28Z","lastTransitionTime":"2025-09-30T13:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.116194 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:57:28 crc kubenswrapper[4840]: E0930 13:57:28.116343 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.183050 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.183082 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.183092 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.183109 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.183120 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:28Z","lastTransitionTime":"2025-09-30T13:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.286096 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.286137 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.286147 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.286161 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.286172 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:28Z","lastTransitionTime":"2025-09-30T13:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.388609 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.388662 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.388687 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.388717 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.388728 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:28Z","lastTransitionTime":"2025-09-30T13:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.491331 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.491373 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.491383 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.491400 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.491412 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:28Z","lastTransitionTime":"2025-09-30T13:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.593471 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.593508 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.593518 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.593534 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.593544 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:28Z","lastTransitionTime":"2025-09-30T13:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.696964 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.697044 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.697073 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.697102 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.697124 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:28Z","lastTransitionTime":"2025-09-30T13:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.799437 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.799472 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.799482 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.799498 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.799509 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:28Z","lastTransitionTime":"2025-09-30T13:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.902827 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.902888 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.902904 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.902926 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:28 crc kubenswrapper[4840]: I0930 13:57:28.902942 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:28Z","lastTransitionTime":"2025-09-30T13:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.006968 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.007041 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.007053 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.007074 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.007087 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:29Z","lastTransitionTime":"2025-09-30T13:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.110609 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.110664 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.110678 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.110709 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.110730 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:29Z","lastTransitionTime":"2025-09-30T13:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.116020 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.116120 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.116020 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:57:29 crc kubenswrapper[4840]: E0930 13:57:29.116227 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:57:29 crc kubenswrapper[4840]: E0930 13:57:29.116451 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:57:29 crc kubenswrapper[4840]: E0930 13:57:29.116489 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.213832 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.213873 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.213884 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.213902 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.213912 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:29Z","lastTransitionTime":"2025-09-30T13:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.316798 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.316835 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.316843 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.316855 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.316864 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:29Z","lastTransitionTime":"2025-09-30T13:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.421474 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.421901 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.421925 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.421951 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.421965 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:29Z","lastTransitionTime":"2025-09-30T13:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.525130 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.525201 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.525213 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.525231 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.525244 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:29Z","lastTransitionTime":"2025-09-30T13:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.539529 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bwvl2_cbe233ee-1ea7-433e-a53a-e4a668f739ee/kube-multus/0.log" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.539603 4840 generic.go:334] "Generic (PLEG): container finished" podID="cbe233ee-1ea7-433e-a53a-e4a668f739ee" containerID="1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a" exitCode=1 Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.539649 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bwvl2" event={"ID":"cbe233ee-1ea7-433e-a53a-e4a668f739ee","Type":"ContainerDied","Data":"1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a"} Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.540188 4840 scope.go:117] "RemoveContainer" containerID="1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.560860 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8d388c29baf0ed22d5c9240f2585e4ca133de265227af6c802a169da6650f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:29Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.575889 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b16edc8f720d4296d7fe50fe407f5983ab975818ec6954d1f0e738e5596624a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:29Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.587952 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:29Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.600422 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c2d8050-9531-4413-9d18-14bf3be65acf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34a542c60726f160b423c3e546ff4769a4a7e4a2441b5a09099e4d8c688286c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b5244dfa04a1f1804ef64dc65030d5eaedf679849eec6a9f0731e6cf147f83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7nvs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:29Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.613006 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gp5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1491f559-bc12-4afd-a40c-4eaa40d920a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gp5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:29Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.626263 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49c5addb-0d04-4c3d-b10e-a07d5fec55da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cf852cfa87ef328d172369c26197a9e6a58fe26fa3113469d9275b2e78c30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a08b53dbed5f3a033d2838519733a2a4109ac5a954d9f5fabd0762b2823b090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3f57df6d22c47163751c297aff0ca1dc2bdbf150b9e385e72acdefde92910c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08be626d75cb6bf53342c4ccab7abef9374735cd29d4afb7c7b5b62715db5d22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08be626d75cb6bf53342c4ccab7abef9374735cd29d4afb7c7b5b62715db5d22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:29Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.627643 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.627671 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.627683 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.627696 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.627710 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:29Z","lastTransitionTime":"2025-09-30T13:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.639738 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f8cd5bd-4a7f-4a5b-ab00-0cc64de9de94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cba320f443ed90a4e755228b663a1da3096e7ea8e068671d1f6273544b6ef60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c507106f39ae2e8acd4268c4e8f28ffd88a0965e69544d8d8a8830e83eceaf9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d62fbefb67a24fedc5f13652596c76f5b288fec58d79b5209f49c79db139d79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61317d26f40a4361a289779bb97b235525b8dddae383f63abfdcab634b633479\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\"\\\\nI0930 13:56:34.911386 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 13:56:34.920195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 13:56:34.920220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 13:56:34.920240 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 13:56:34.920248 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 13:56:34.927953 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 13:56:34.927977 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927984 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 13:56:34.927992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 13:56:34.927996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 13:56:34.928001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 13:56:34.928159 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0930 13:56:34.930945 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759240579\\\\\\\\\\\\\\\" (2025-09-30 13:56:19 +0000 UTC to 2025-10-30 13:56:20 +0000 UTC (now=2025-09-30 13:56:34.930900435 +0000 UTC))\\\\\\\"\\\\nF0930 13:56:34.930995 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1435e2e88d45e5561ff5863095d0f953ef1478d19d87691ab226cad7ae148f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:29Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.651134 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:29Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.673828 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84052a424fd6279d70b54902ebdcf430c8aa53b02406b52fc261db9d4d88ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27a94b18137fa991459484406b1c5bd2b2dbb52f527d5f13156ba8e5b8d0146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc4e5d3625aa2d5f94307cedddf72cc48f17f327a044501144e35422d670d555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9549ce85242f1f19b3a52e603caec58bb4c3e2449eb50d8c4525f56dfc93e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a364a13c773fe00ee4e022a286821013d3a681acc16567e66b1c33c7e18ee3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79f4e921ab2d7e8370fe83552c960147cbbd744f03bca3e390813ca3d021aa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c79f4e921ab2d7e8370fe83552c960147cbbd744f03bca3e390813ca3d021aa3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T13:57:09Z\\\",\\\"message\\\":\\\"irewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 13:57:09.078226 6498 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0930 13:57:09.078250 6498 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0930 13:57:09.078275 6498 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0930 13:57:09.078279 6498 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0930 13:57:09.078296 6498 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0930 13:57:09.078313 6498 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0930 13:57:09.078326 6498 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0930 13:57:09.078344 6498 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0930 13:57:09.078531 6498 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0930 13:57:09.078543 6498 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0930 13:57:09.078563 6498 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0930 13:57:09.078571 6498 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 13:57:09.078579 6498 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0930 13:57:09.078585 6498 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0930 13:57:09.078708 6498 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:57:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-kfx69_openshift-ovn-kubernetes(2ff48c28-d076-46e8-a93f-9630989f81e8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:29Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.691858 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T13:57:29Z\\\",\\\"message\\\":\\\"2025-09-30T13:56:43+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_afea8585-8b55-4126-ad48-e6082b820f13\\\\n2025-09-30T13:56:43+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_afea8585-8b55-4126-ad48-e6082b820f13 to /host/opt/cni/bin/\\\\n2025-09-30T13:56:44Z [verbose] multus-daemon started\\\\n2025-09-30T13:56:44Z [verbose] Readiness Indicator file check\\\\n2025-09-30T13:57:29Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:29Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.706390 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:29Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.716974 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pj2zd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4de996b8-bda8-4142-b5e0-04ed9ae1f327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7286e2d685d6573993e2fa000cb9e08d85c6aa820af507327be52ca50fe58c7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pj2zd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:29Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.728217 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:29Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.729477 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.729502 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.729511 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.729526 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.729537 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:29Z","lastTransitionTime":"2025-09-30T13:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.740866 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:29Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.752786 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:29Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.764837 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:29Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.832153 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.832184 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.832192 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.832206 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.832215 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:29Z","lastTransitionTime":"2025-09-30T13:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.934730 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.934784 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.934801 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.934822 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:29 crc kubenswrapper[4840]: I0930 13:57:29.934836 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:29Z","lastTransitionTime":"2025-09-30T13:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.038536 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.038601 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.038612 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.038636 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.038649 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:30Z","lastTransitionTime":"2025-09-30T13:57:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.115957 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:57:30 crc kubenswrapper[4840]: E0930 13:57:30.116591 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.136051 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.141504 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.141578 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.141591 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.141613 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.141628 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:30Z","lastTransitionTime":"2025-09-30T13:57:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.244370 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.244413 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.244424 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.244440 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.244451 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:30Z","lastTransitionTime":"2025-09-30T13:57:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.347241 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.347294 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.347313 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.347332 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.347343 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:30Z","lastTransitionTime":"2025-09-30T13:57:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.450071 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.450164 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.450188 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.450216 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.450235 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:30Z","lastTransitionTime":"2025-09-30T13:57:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.546489 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bwvl2_cbe233ee-1ea7-433e-a53a-e4a668f739ee/kube-multus/0.log" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.546656 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bwvl2" event={"ID":"cbe233ee-1ea7-433e-a53a-e4a668f739ee","Type":"ContainerStarted","Data":"a3eb5795ae35b00eba9462e95fdba89c2dcf0442843c32ec05edd710dfe507fe"} Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.551907 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.551992 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.552027 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.552062 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.552090 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:30Z","lastTransitionTime":"2025-09-30T13:57:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.560180 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49c5addb-0d04-4c3d-b10e-a07d5fec55da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cf852cfa87ef328d172369c26197a9e6a58fe26fa3113469d9275b2e78c30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a08b53dbed5f3a033d2838519733a2a4109ac5a954d9f5fabd0762b2823b090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3f57df6d22c47163751c297aff0ca1dc2bdbf150b9e385e72acdefde92910c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08be626d75cb6bf53342c4ccab7abef9374735cd29d4afb7c7b5b62715db5d22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08be626d75cb6bf53342c4ccab7abef9374735cd29d4afb7c7b5b62715db5d22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:30Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.573434 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f8cd5bd-4a7f-4a5b-ab00-0cc64de9de94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cba320f443ed90a4e755228b663a1da3096e7ea8e068671d1f6273544b6ef60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c507106f39ae2e8acd4268c4e8f28ffd88a0965e69544d8d8a8830e83eceaf9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d62fbefb67a24fedc5f13652596c76f5b288fec58d79b5209f49c79db139d79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61317d26f40a4361a289779bb97b235525b8dddae383f63abfdcab634b633479\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\"\\\\nI0930 13:56:34.911386 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 13:56:34.920195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 13:56:34.920220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 13:56:34.920240 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 13:56:34.920248 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 13:56:34.927953 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 13:56:34.927977 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927984 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 13:56:34.927992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 13:56:34.927996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 13:56:34.928001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 13:56:34.928159 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0930 13:56:34.930945 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759240579\\\\\\\\\\\\\\\" (2025-09-30 13:56:19 +0000 UTC to 2025-10-30 13:56:20 +0000 UTC (now=2025-09-30 13:56:34.930900435 +0000 UTC))\\\\\\\"\\\\nF0930 13:56:34.930995 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1435e2e88d45e5561ff5863095d0f953ef1478d19d87691ab226cad7ae148f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:30Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.585016 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:30Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.605816 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8d388c29baf0ed22d5c9240f2585e4ca133de265227af6c802a169da6650f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:30Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.617837 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b16edc8f720d4296d7fe50fe407f5983ab975818ec6954d1f0e738e5596624a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:30Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.629764 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:30Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.640018 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c2d8050-9531-4413-9d18-14bf3be65acf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34a542c60726f160b423c3e546ff4769a4a7e4a2441b5a09099e4d8c688286c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b5244dfa04a1f1804ef64dc65030d5eaedf679849eec6a9f0731e6cf147f83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7nvs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:30Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.649653 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gp5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1491f559-bc12-4afd-a40c-4eaa40d920a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gp5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:30Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.654149 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.654196 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.654211 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.654233 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.654248 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:30Z","lastTransitionTime":"2025-09-30T13:57:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.666759 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84052a424fd6279d70b54902ebdcf430c8aa53b02406b52fc261db9d4d88ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27a94b18137fa991459484406b1c5bd2b2dbb52f527d5f13156ba8e5b8d0146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc4e5d3625aa2d5f94307cedddf72cc48f17f327a044501144e35422d670d555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9549ce85242f1f19b3a52e603caec58bb4c3e2449eb50d8c4525f56dfc93e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a364a13c773fe00ee4e022a286821013d3a681acc16567e66b1c33c7e18ee3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79f4e921ab2d7e8370fe83552c960147cbbd744f03bca3e390813ca3d021aa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c79f4e921ab2d7e8370fe83552c960147cbbd744f03bca3e390813ca3d021aa3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T13:57:09Z\\\",\\\"message\\\":\\\"irewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 13:57:09.078226 6498 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0930 13:57:09.078250 6498 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0930 13:57:09.078275 6498 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0930 13:57:09.078279 6498 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0930 13:57:09.078296 6498 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0930 13:57:09.078313 6498 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0930 13:57:09.078326 6498 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0930 13:57:09.078344 6498 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0930 13:57:09.078531 6498 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0930 13:57:09.078543 6498 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0930 13:57:09.078563 6498 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0930 13:57:09.078571 6498 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 13:57:09.078579 6498 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0930 13:57:09.078585 6498 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0930 13:57:09.078708 6498 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:57:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-kfx69_openshift-ovn-kubernetes(2ff48c28-d076-46e8-a93f-9630989f81e8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:30Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.679878 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3eb5795ae35b00eba9462e95fdba89c2dcf0442843c32ec05edd710dfe507fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T13:57:29Z\\\",\\\"message\\\":\\\"2025-09-30T13:56:43+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_afea8585-8b55-4126-ad48-e6082b820f13\\\\n2025-09-30T13:56:43+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_afea8585-8b55-4126-ad48-e6082b820f13 to /host/opt/cni/bin/\\\\n2025-09-30T13:56:44Z [verbose] multus-daemon started\\\\n2025-09-30T13:56:44Z [verbose] Readiness Indicator file check\\\\n2025-09-30T13:57:29Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:30Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.700229 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a4dccdc-e5ba-4cfa-ac4b-e34a27c7368f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15be3c94974cfafcf69a74ef1ff8f6b9da74840805ceafb96aeb903dbf90fd58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://987b6a297870d2898fe2b5df90b15166fb192e10d232fa9bae686ebb7f4180aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a3cd4212aed96f1cad501a5c62065b9c1a289c03afbec269ca193c12434c56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15291f77e6b50679df8feef2f536b5f5745d6f841372770d647831006d8eb0af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d68b17f8a9e41caf54d2f8e2ae775c33d653299e8fe157de0fa4596e1339682e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5a84f046af7784b7ef2afb1b87d108573cd76d20f1b74077166f4dd7594f0b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5a84f046af7784b7ef2afb1b87d108573cd76d20f1b74077166f4dd7594f0b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c2c3a885cd937e387b3ba820ef501fe0c318a277d2353aa6664b3674faf79de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c2c3a885cd937e387b3ba820ef501fe0c318a277d2353aa6664b3674faf79de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f335b20cc833d3278d09697f85b1abd85732aefd80429fb1866eca5565ef98e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f335b20cc833d3278d09697f85b1abd85732aefd80429fb1866eca5565ef98e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:30Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.713318 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:30Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.724758 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:30Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.736637 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:30Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.747338 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pj2zd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4de996b8-bda8-4142-b5e0-04ed9ae1f327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7286e2d685d6573993e2fa000cb9e08d85c6aa820af507327be52ca50fe58c7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pj2zd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:30Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.756350 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.756393 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.756404 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.756421 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.756433 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:30Z","lastTransitionTime":"2025-09-30T13:57:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.759283 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:30Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.770412 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:30Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.858329 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.858392 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.858403 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.858419 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.858431 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:30Z","lastTransitionTime":"2025-09-30T13:57:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.961731 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.961807 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.961828 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.961857 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:30 crc kubenswrapper[4840]: I0930 13:57:30.961880 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:30Z","lastTransitionTime":"2025-09-30T13:57:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.065131 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.065173 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.065182 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.065198 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.065207 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:31Z","lastTransitionTime":"2025-09-30T13:57:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.116306 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.116450 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:57:31 crc kubenswrapper[4840]: E0930 13:57:31.116581 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.116663 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:57:31 crc kubenswrapper[4840]: E0930 13:57:31.116677 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:57:31 crc kubenswrapper[4840]: E0930 13:57:31.116803 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.167575 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.167636 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.167669 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.167689 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.167701 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:31Z","lastTransitionTime":"2025-09-30T13:57:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.269837 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.269876 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.269886 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.269901 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.269912 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:31Z","lastTransitionTime":"2025-09-30T13:57:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.372609 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.372644 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.372651 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.372666 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.372675 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:31Z","lastTransitionTime":"2025-09-30T13:57:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.475338 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.475393 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.475408 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.475430 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.475445 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:31Z","lastTransitionTime":"2025-09-30T13:57:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.578342 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.578423 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.578446 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.578515 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.578540 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:31Z","lastTransitionTime":"2025-09-30T13:57:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.681352 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.681400 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.681412 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.681434 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.681446 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:31Z","lastTransitionTime":"2025-09-30T13:57:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.784796 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.784856 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.784867 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.784888 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.784901 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:31Z","lastTransitionTime":"2025-09-30T13:57:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.887621 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.887666 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.887678 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.887698 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.887712 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:31Z","lastTransitionTime":"2025-09-30T13:57:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.990201 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.990232 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.990239 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.990254 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:31 crc kubenswrapper[4840]: I0930 13:57:31.990263 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:31Z","lastTransitionTime":"2025-09-30T13:57:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.092730 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.092765 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.092773 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.092786 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.092798 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:32Z","lastTransitionTime":"2025-09-30T13:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.116087 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:57:32 crc kubenswrapper[4840]: E0930 13:57:32.116264 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.134748 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b16edc8f720d4296d7fe50fe407f5983ab975818ec6954d1f0e738e5596624a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:32Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.145521 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:32Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.158142 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c2d8050-9531-4413-9d18-14bf3be65acf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34a542c60726f160b423c3e546ff4769a4a7e4a2441b5a09099e4d8c688286c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b5244dfa04a1f1804ef64dc65030d5eaedf679849eec6a9f0731e6cf147f83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7nvs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:32Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.171338 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gp5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1491f559-bc12-4afd-a40c-4eaa40d920a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gp5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:32Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.183963 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49c5addb-0d04-4c3d-b10e-a07d5fec55da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cf852cfa87ef328d172369c26197a9e6a58fe26fa3113469d9275b2e78c30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a08b53dbed5f3a033d2838519733a2a4109ac5a954d9f5fabd0762b2823b090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3f57df6d22c47163751c297aff0ca1dc2bdbf150b9e385e72acdefde92910c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08be626d75cb6bf53342c4ccab7abef9374735cd29d4afb7c7b5b62715db5d22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08be626d75cb6bf53342c4ccab7abef9374735cd29d4afb7c7b5b62715db5d22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:32Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.195618 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.195899 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.195978 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.196052 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.196122 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:32Z","lastTransitionTime":"2025-09-30T13:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.198393 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f8cd5bd-4a7f-4a5b-ab00-0cc64de9de94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cba320f443ed90a4e755228b663a1da3096e7ea8e068671d1f6273544b6ef60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c507106f39ae2e8acd4268c4e8f28ffd88a0965e69544d8d8a8830e83eceaf9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d62fbefb67a24fedc5f13652596c76f5b288fec58d79b5209f49c79db139d79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61317d26f40a4361a289779bb97b235525b8dddae383f63abfdcab634b633479\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\"\\\\nI0930 13:56:34.911386 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 13:56:34.920195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 13:56:34.920220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 13:56:34.920240 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 13:56:34.920248 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 13:56:34.927953 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 13:56:34.927977 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927984 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 13:56:34.927992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 13:56:34.927996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 13:56:34.928001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 13:56:34.928159 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0930 13:56:34.930945 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759240579\\\\\\\\\\\\\\\" (2025-09-30 13:56:19 +0000 UTC to 2025-10-30 13:56:20 +0000 UTC (now=2025-09-30 13:56:34.930900435 +0000 UTC))\\\\\\\"\\\\nF0930 13:56:34.930995 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1435e2e88d45e5561ff5863095d0f953ef1478d19d87691ab226cad7ae148f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:32Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.212597 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:32Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.230471 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8d388c29baf0ed22d5c9240f2585e4ca133de265227af6c802a169da6650f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:32Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.258757 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84052a424fd6279d70b54902ebdcf430c8aa53b02406b52fc261db9d4d88ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27a94b18137fa991459484406b1c5bd2b2dbb52f527d5f13156ba8e5b8d0146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc4e5d3625aa2d5f94307cedddf72cc48f17f327a044501144e35422d670d555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9549ce85242f1f19b3a52e603caec58bb4c3e2449eb50d8c4525f56dfc93e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a364a13c773fe00ee4e022a286821013d3a681acc16567e66b1c33c7e18ee3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79f4e921ab2d7e8370fe83552c960147cbbd744f03bca3e390813ca3d021aa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c79f4e921ab2d7e8370fe83552c960147cbbd744f03bca3e390813ca3d021aa3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T13:57:09Z\\\",\\\"message\\\":\\\"irewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 13:57:09.078226 6498 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0930 13:57:09.078250 6498 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0930 13:57:09.078275 6498 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0930 13:57:09.078279 6498 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0930 13:57:09.078296 6498 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0930 13:57:09.078313 6498 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0930 13:57:09.078326 6498 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0930 13:57:09.078344 6498 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0930 13:57:09.078531 6498 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0930 13:57:09.078543 6498 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0930 13:57:09.078563 6498 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0930 13:57:09.078571 6498 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 13:57:09.078579 6498 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0930 13:57:09.078585 6498 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0930 13:57:09.078708 6498 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:57:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-kfx69_openshift-ovn-kubernetes(2ff48c28-d076-46e8-a93f-9630989f81e8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:32Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.272377 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3eb5795ae35b00eba9462e95fdba89c2dcf0442843c32ec05edd710dfe507fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T13:57:29Z\\\",\\\"message\\\":\\\"2025-09-30T13:56:43+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_afea8585-8b55-4126-ad48-e6082b820f13\\\\n2025-09-30T13:56:43+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_afea8585-8b55-4126-ad48-e6082b820f13 to /host/opt/cni/bin/\\\\n2025-09-30T13:56:44Z [verbose] multus-daemon started\\\\n2025-09-30T13:56:44Z [verbose] Readiness Indicator file check\\\\n2025-09-30T13:57:29Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:32Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.282897 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pj2zd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4de996b8-bda8-4142-b5e0-04ed9ae1f327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7286e2d685d6573993e2fa000cb9e08d85c6aa820af507327be52ca50fe58c7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pj2zd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:32Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.297996 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.298025 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.298035 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.298048 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.298060 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:32Z","lastTransitionTime":"2025-09-30T13:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.302983 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a4dccdc-e5ba-4cfa-ac4b-e34a27c7368f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15be3c94974cfafcf69a74ef1ff8f6b9da74840805ceafb96aeb903dbf90fd58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://987b6a297870d2898fe2b5df90b15166fb192e10d232fa9bae686ebb7f4180aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a3cd4212aed96f1cad501a5c62065b9c1a289c03afbec269ca193c12434c56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15291f77e6b50679df8feef2f536b5f5745d6f841372770d647831006d8eb0af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d68b17f8a9e41caf54d2f8e2ae775c33d653299e8fe157de0fa4596e1339682e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5a84f046af7784b7ef2afb1b87d108573cd76d20f1b74077166f4dd7594f0b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5a84f046af7784b7ef2afb1b87d108573cd76d20f1b74077166f4dd7594f0b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c2c3a885cd937e387b3ba820ef501fe0c318a277d2353aa6664b3674faf79de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c2c3a885cd937e387b3ba820ef501fe0c318a277d2353aa6664b3674faf79de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f335b20cc833d3278d09697f85b1abd85732aefd80429fb1866eca5565ef98e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f335b20cc833d3278d09697f85b1abd85732aefd80429fb1866eca5565ef98e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:32Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.321765 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:32Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.334971 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:32Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.350512 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:32Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.364248 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:32Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.379120 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:32Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.400005 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.400042 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.400051 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.400065 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.400074 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:32Z","lastTransitionTime":"2025-09-30T13:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.502043 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.502100 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.502115 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.502141 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.502156 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:32Z","lastTransitionTime":"2025-09-30T13:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.605175 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.605240 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.605258 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.605282 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.605301 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:32Z","lastTransitionTime":"2025-09-30T13:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.708198 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.708278 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.708300 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.708329 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.708350 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:32Z","lastTransitionTime":"2025-09-30T13:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.810836 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.810875 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.810886 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.810905 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.810915 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:32Z","lastTransitionTime":"2025-09-30T13:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.914089 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.914159 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.914181 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.914211 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:32 crc kubenswrapper[4840]: I0930 13:57:32.914235 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:32Z","lastTransitionTime":"2025-09-30T13:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.017305 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.017357 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.017366 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.017381 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.017389 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:33Z","lastTransitionTime":"2025-09-30T13:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.115873 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.115936 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.115873 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:57:33 crc kubenswrapper[4840]: E0930 13:57:33.116012 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:57:33 crc kubenswrapper[4840]: E0930 13:57:33.116181 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:57:33 crc kubenswrapper[4840]: E0930 13:57:33.116297 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.120209 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.120268 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.120279 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.120292 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.120304 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:33Z","lastTransitionTime":"2025-09-30T13:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.223085 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.223143 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.223154 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.223170 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.223181 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:33Z","lastTransitionTime":"2025-09-30T13:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.326281 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.326347 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.326367 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.326393 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.326417 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:33Z","lastTransitionTime":"2025-09-30T13:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.428931 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.428976 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.428986 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.429002 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.429014 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:33Z","lastTransitionTime":"2025-09-30T13:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.531902 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.531970 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.531987 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.532011 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.532029 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:33Z","lastTransitionTime":"2025-09-30T13:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.634318 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.634373 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.634389 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.634408 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.634423 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:33Z","lastTransitionTime":"2025-09-30T13:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.737383 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.737458 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.737497 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.737525 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.737546 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:33Z","lastTransitionTime":"2025-09-30T13:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.840838 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.840875 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.840889 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.840902 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.840912 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:33Z","lastTransitionTime":"2025-09-30T13:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.943914 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.943968 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.943984 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.944006 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:33 crc kubenswrapper[4840]: I0930 13:57:33.944023 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:33Z","lastTransitionTime":"2025-09-30T13:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.046168 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.046218 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.046227 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.046240 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.046248 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:34Z","lastTransitionTime":"2025-09-30T13:57:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.116215 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:57:34 crc kubenswrapper[4840]: E0930 13:57:34.116399 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.148596 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.148642 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.148651 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.148666 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.148676 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:34Z","lastTransitionTime":"2025-09-30T13:57:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.251917 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.251965 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.251977 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.251993 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.252004 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:34Z","lastTransitionTime":"2025-09-30T13:57:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.357157 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.357440 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.357512 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.357597 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.357667 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:34Z","lastTransitionTime":"2025-09-30T13:57:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.461234 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.461305 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.461325 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.461351 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.461369 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:34Z","lastTransitionTime":"2025-09-30T13:57:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.564517 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.564604 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.564621 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.564643 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.564662 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:34Z","lastTransitionTime":"2025-09-30T13:57:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.667272 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.667346 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.667358 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.667378 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.667391 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:34Z","lastTransitionTime":"2025-09-30T13:57:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.769846 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.769885 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.769898 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.769915 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.769926 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:34Z","lastTransitionTime":"2025-09-30T13:57:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.873110 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.873152 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.873160 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.873179 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.873188 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:34Z","lastTransitionTime":"2025-09-30T13:57:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.975027 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.975060 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.975068 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.975082 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:34 crc kubenswrapper[4840]: I0930 13:57:34.975108 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:34Z","lastTransitionTime":"2025-09-30T13:57:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.078569 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.078619 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.078630 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.078650 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.078662 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:35Z","lastTransitionTime":"2025-09-30T13:57:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.115566 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:57:35 crc kubenswrapper[4840]: E0930 13:57:35.115737 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.115825 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.115944 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:57:35 crc kubenswrapper[4840]: E0930 13:57:35.115981 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:57:35 crc kubenswrapper[4840]: E0930 13:57:35.116168 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.181295 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.181376 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.181400 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.181434 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.181450 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:35Z","lastTransitionTime":"2025-09-30T13:57:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.283341 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.283375 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.283384 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.283398 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.283408 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:35Z","lastTransitionTime":"2025-09-30T13:57:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.385882 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.385921 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.385932 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.385947 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.385958 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:35Z","lastTransitionTime":"2025-09-30T13:57:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.488773 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.488848 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.488865 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.488892 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.488910 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:35Z","lastTransitionTime":"2025-09-30T13:57:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.590697 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.590733 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.590744 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.590759 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.590769 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:35Z","lastTransitionTime":"2025-09-30T13:57:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.693192 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.693237 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.693248 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.693264 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.693275 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:35Z","lastTransitionTime":"2025-09-30T13:57:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.795902 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.795953 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.795966 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.795983 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.795993 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:35Z","lastTransitionTime":"2025-09-30T13:57:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.901106 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.901174 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.901192 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.901218 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:35 crc kubenswrapper[4840]: I0930 13:57:35.901236 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:35Z","lastTransitionTime":"2025-09-30T13:57:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.003649 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.003679 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.003687 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.003699 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.003708 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:36Z","lastTransitionTime":"2025-09-30T13:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.106879 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.106950 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.106967 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.106993 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.107009 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:36Z","lastTransitionTime":"2025-09-30T13:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.116439 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:57:36 crc kubenswrapper[4840]: E0930 13:57:36.116642 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.209287 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.209348 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.209364 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.209386 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.209408 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:36Z","lastTransitionTime":"2025-09-30T13:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.312966 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.313026 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.313043 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.313082 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.313116 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:36Z","lastTransitionTime":"2025-09-30T13:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.417139 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.417206 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.417223 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.417251 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.417267 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:36Z","lastTransitionTime":"2025-09-30T13:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.519805 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.519863 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.519880 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.519898 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.519910 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:36Z","lastTransitionTime":"2025-09-30T13:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.622866 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.622942 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.622967 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.622996 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.623014 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:36Z","lastTransitionTime":"2025-09-30T13:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.733594 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.733658 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.733673 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.733697 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.733713 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:36Z","lastTransitionTime":"2025-09-30T13:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.836791 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.836861 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.836874 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.836891 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.836903 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:36Z","lastTransitionTime":"2025-09-30T13:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.870989 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.871086 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.871101 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.871119 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.871131 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:36Z","lastTransitionTime":"2025-09-30T13:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:36 crc kubenswrapper[4840]: E0930 13:57:36.889693 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5\\\",\\\"systemUUID\\\":\\\"f41e0351-4eda-411a-9a99-43a1952f3d34\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:36Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.893543 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.893636 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.895210 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.895288 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.895313 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:36Z","lastTransitionTime":"2025-09-30T13:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:36 crc kubenswrapper[4840]: E0930 13:57:36.915345 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5\\\",\\\"systemUUID\\\":\\\"f41e0351-4eda-411a-9a99-43a1952f3d34\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:36Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.920275 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.920347 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.920360 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.920379 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.920393 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:36Z","lastTransitionTime":"2025-09-30T13:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:36 crc kubenswrapper[4840]: E0930 13:57:36.933253 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5\\\",\\\"systemUUID\\\":\\\"f41e0351-4eda-411a-9a99-43a1952f3d34\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:36Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.939148 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.939186 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.939198 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.939220 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.939235 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:36Z","lastTransitionTime":"2025-09-30T13:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:36 crc kubenswrapper[4840]: E0930 13:57:36.953643 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5\\\",\\\"systemUUID\\\":\\\"f41e0351-4eda-411a-9a99-43a1952f3d34\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:36Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.957986 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.958050 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.958065 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.958089 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.958102 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:36Z","lastTransitionTime":"2025-09-30T13:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:36 crc kubenswrapper[4840]: E0930 13:57:36.973785 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5\\\",\\\"systemUUID\\\":\\\"f41e0351-4eda-411a-9a99-43a1952f3d34\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:36Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:36 crc kubenswrapper[4840]: E0930 13:57:36.973924 4840 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.976033 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.976080 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.976093 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.976112 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:36 crc kubenswrapper[4840]: I0930 13:57:36.976127 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:36Z","lastTransitionTime":"2025-09-30T13:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.078982 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.079023 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.079040 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.079059 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.079073 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:37Z","lastTransitionTime":"2025-09-30T13:57:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.115780 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.115862 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.115899 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:57:37 crc kubenswrapper[4840]: E0930 13:57:37.116103 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:57:37 crc kubenswrapper[4840]: E0930 13:57:37.116279 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:57:37 crc kubenswrapper[4840]: E0930 13:57:37.116434 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.182945 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.183028 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.183052 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.183081 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.183100 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:37Z","lastTransitionTime":"2025-09-30T13:57:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.285306 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.285352 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.285363 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.285378 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.285392 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:37Z","lastTransitionTime":"2025-09-30T13:57:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.387957 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.387993 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.388001 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.388017 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.388027 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:37Z","lastTransitionTime":"2025-09-30T13:57:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.490783 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.490865 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.490882 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.490904 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.490921 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:37Z","lastTransitionTime":"2025-09-30T13:57:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.593983 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.594379 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.594632 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.594836 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.594998 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:37Z","lastTransitionTime":"2025-09-30T13:57:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.698331 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.699514 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.699769 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.700001 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.700219 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:37Z","lastTransitionTime":"2025-09-30T13:57:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.803166 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.803427 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.803442 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.803456 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.803466 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:37Z","lastTransitionTime":"2025-09-30T13:57:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.906149 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.906182 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.906194 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.906210 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:37 crc kubenswrapper[4840]: I0930 13:57:37.906221 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:37Z","lastTransitionTime":"2025-09-30T13:57:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.008570 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.008604 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.008612 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.008624 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.008634 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:38Z","lastTransitionTime":"2025-09-30T13:57:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.111694 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.111745 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.111757 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.111774 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.111786 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:38Z","lastTransitionTime":"2025-09-30T13:57:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.117131 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:57:38 crc kubenswrapper[4840]: E0930 13:57:38.117310 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.117718 4840 scope.go:117] "RemoveContainer" containerID="c79f4e921ab2d7e8370fe83552c960147cbbd744f03bca3e390813ca3d021aa3" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.214410 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.214763 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.214774 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.214790 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.214807 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:38Z","lastTransitionTime":"2025-09-30T13:57:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.317320 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.317363 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.317375 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.317389 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.317399 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:38Z","lastTransitionTime":"2025-09-30T13:57:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.420608 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.420668 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.420683 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.420706 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.420722 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:38Z","lastTransitionTime":"2025-09-30T13:57:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.522857 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.522926 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.522937 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.522956 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.522968 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:38Z","lastTransitionTime":"2025-09-30T13:57:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.576781 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kfx69_2ff48c28-d076-46e8-a93f-9630989f81e8/ovnkube-controller/2.log" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.579381 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" event={"ID":"2ff48c28-d076-46e8-a93f-9630989f81e8","Type":"ContainerStarted","Data":"109628b3a230ce972e436d712479c6f8071e34a983f83d79555ffea902ca2e55"} Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.579813 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.593903 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:38Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.608003 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:38Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.626060 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.626184 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.626201 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.626225 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.626255 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:38Z","lastTransitionTime":"2025-09-30T13:57:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.639201 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8d388c29baf0ed22d5c9240f2585e4ca133de265227af6c802a169da6650f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:38Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.658460 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b16edc8f720d4296d7fe50fe407f5983ab975818ec6954d1f0e738e5596624a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:38Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.669120 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:38Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.680146 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c2d8050-9531-4413-9d18-14bf3be65acf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34a542c60726f160b423c3e546ff4769a4a7e4a2441b5a09099e4d8c688286c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b5244dfa04a1f1804ef64dc65030d5eaedf679849eec6a9f0731e6cf147f83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7nvs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:38Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.694760 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gp5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1491f559-bc12-4afd-a40c-4eaa40d920a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gp5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:38Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.707686 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49c5addb-0d04-4c3d-b10e-a07d5fec55da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cf852cfa87ef328d172369c26197a9e6a58fe26fa3113469d9275b2e78c30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a08b53dbed5f3a033d2838519733a2a4109ac5a954d9f5fabd0762b2823b090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3f57df6d22c47163751c297aff0ca1dc2bdbf150b9e385e72acdefde92910c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08be626d75cb6bf53342c4ccab7abef9374735cd29d4afb7c7b5b62715db5d22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08be626d75cb6bf53342c4ccab7abef9374735cd29d4afb7c7b5b62715db5d22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:38Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.720234 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f8cd5bd-4a7f-4a5b-ab00-0cc64de9de94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cba320f443ed90a4e755228b663a1da3096e7ea8e068671d1f6273544b6ef60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c507106f39ae2e8acd4268c4e8f28ffd88a0965e69544d8d8a8830e83eceaf9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d62fbefb67a24fedc5f13652596c76f5b288fec58d79b5209f49c79db139d79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61317d26f40a4361a289779bb97b235525b8dddae383f63abfdcab634b633479\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\"\\\\nI0930 13:56:34.911386 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 13:56:34.920195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 13:56:34.920220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 13:56:34.920240 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 13:56:34.920248 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 13:56:34.927953 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 13:56:34.927977 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927984 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 13:56:34.927992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 13:56:34.927996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 13:56:34.928001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 13:56:34.928159 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0930 13:56:34.930945 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759240579\\\\\\\\\\\\\\\" (2025-09-30 13:56:19 +0000 UTC to 2025-10-30 13:56:20 +0000 UTC (now=2025-09-30 13:56:34.930900435 +0000 UTC))\\\\\\\"\\\\nF0930 13:56:34.930995 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1435e2e88d45e5561ff5863095d0f953ef1478d19d87691ab226cad7ae148f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:38Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.729350 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.729383 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.729391 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.729405 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.729414 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:38Z","lastTransitionTime":"2025-09-30T13:57:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.731470 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:38Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.748497 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84052a424fd6279d70b54902ebdcf430c8aa53b02406b52fc261db9d4d88ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27a94b18137fa991459484406b1c5bd2b2dbb52f527d5f13156ba8e5b8d0146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc4e5d3625aa2d5f94307cedddf72cc48f17f327a044501144e35422d670d555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9549ce85242f1f19b3a52e603caec58bb4c3e2449eb50d8c4525f56dfc93e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a364a13c773fe00ee4e022a286821013d3a681acc16567e66b1c33c7e18ee3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://109628b3a230ce972e436d712479c6f8071e34a983f83d79555ffea902ca2e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c79f4e921ab2d7e8370fe83552c960147cbbd744f03bca3e390813ca3d021aa3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T13:57:09Z\\\",\\\"message\\\":\\\"irewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 13:57:09.078226 6498 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0930 13:57:09.078250 6498 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0930 13:57:09.078275 6498 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0930 13:57:09.078279 6498 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0930 13:57:09.078296 6498 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0930 13:57:09.078313 6498 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0930 13:57:09.078326 6498 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0930 13:57:09.078344 6498 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0930 13:57:09.078531 6498 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0930 13:57:09.078543 6498 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0930 13:57:09.078563 6498 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0930 13:57:09.078571 6498 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 13:57:09.078579 6498 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0930 13:57:09.078585 6498 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0930 13:57:09.078708 6498 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:57:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:57:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:38Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.760104 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3eb5795ae35b00eba9462e95fdba89c2dcf0442843c32ec05edd710dfe507fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T13:57:29Z\\\",\\\"message\\\":\\\"2025-09-30T13:56:43+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_afea8585-8b55-4126-ad48-e6082b820f13\\\\n2025-09-30T13:56:43+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_afea8585-8b55-4126-ad48-e6082b820f13 to /host/opt/cni/bin/\\\\n2025-09-30T13:56:44Z [verbose] multus-daemon started\\\\n2025-09-30T13:56:44Z [verbose] Readiness Indicator file check\\\\n2025-09-30T13:57:29Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:38Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.770216 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:38Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.779537 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pj2zd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4de996b8-bda8-4142-b5e0-04ed9ae1f327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7286e2d685d6573993e2fa000cb9e08d85c6aa820af507327be52ca50fe58c7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pj2zd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:38Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.797066 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a4dccdc-e5ba-4cfa-ac4b-e34a27c7368f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15be3c94974cfafcf69a74ef1ff8f6b9da74840805ceafb96aeb903dbf90fd58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://987b6a297870d2898fe2b5df90b15166fb192e10d232fa9bae686ebb7f4180aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a3cd4212aed96f1cad501a5c62065b9c1a289c03afbec269ca193c12434c56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15291f77e6b50679df8feef2f536b5f5745d6f841372770d647831006d8eb0af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d68b17f8a9e41caf54d2f8e2ae775c33d653299e8fe157de0fa4596e1339682e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5a84f046af7784b7ef2afb1b87d108573cd76d20f1b74077166f4dd7594f0b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5a84f046af7784b7ef2afb1b87d108573cd76d20f1b74077166f4dd7594f0b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c2c3a885cd937e387b3ba820ef501fe0c318a277d2353aa6664b3674faf79de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c2c3a885cd937e387b3ba820ef501fe0c318a277d2353aa6664b3674faf79de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f335b20cc833d3278d09697f85b1abd85732aefd80429fb1866eca5565ef98e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f335b20cc833d3278d09697f85b1abd85732aefd80429fb1866eca5565ef98e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:38Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.810717 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:38Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.822782 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:38Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.831237 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.831268 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.831279 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.831293 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.831302 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:38Z","lastTransitionTime":"2025-09-30T13:57:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.933631 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.933675 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.933692 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.933709 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:38 crc kubenswrapper[4840]: I0930 13:57:38.933720 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:38Z","lastTransitionTime":"2025-09-30T13:57:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.036264 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.036327 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.036351 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.036381 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.036406 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:39Z","lastTransitionTime":"2025-09-30T13:57:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.115432 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.115465 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.115474 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:57:39 crc kubenswrapper[4840]: E0930 13:57:39.115644 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:57:39 crc kubenswrapper[4840]: E0930 13:57:39.115748 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:57:39 crc kubenswrapper[4840]: E0930 13:57:39.116014 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.139863 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.139926 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.139948 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.139976 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.139998 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:39Z","lastTransitionTime":"2025-09-30T13:57:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.242935 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.242996 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.243013 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.243039 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.243057 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:39Z","lastTransitionTime":"2025-09-30T13:57:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.346017 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.346053 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.346065 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.346082 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.346094 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:39Z","lastTransitionTime":"2025-09-30T13:57:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.448632 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.448695 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.448714 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.448738 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.448753 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:39Z","lastTransitionTime":"2025-09-30T13:57:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.551892 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.551932 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.551942 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.551956 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.551965 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:39Z","lastTransitionTime":"2025-09-30T13:57:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.584727 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kfx69_2ff48c28-d076-46e8-a93f-9630989f81e8/ovnkube-controller/3.log" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.585533 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kfx69_2ff48c28-d076-46e8-a93f-9630989f81e8/ovnkube-controller/2.log" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.587847 4840 generic.go:334] "Generic (PLEG): container finished" podID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerID="109628b3a230ce972e436d712479c6f8071e34a983f83d79555ffea902ca2e55" exitCode=1 Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.587883 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" event={"ID":"2ff48c28-d076-46e8-a93f-9630989f81e8","Type":"ContainerDied","Data":"109628b3a230ce972e436d712479c6f8071e34a983f83d79555ffea902ca2e55"} Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.587919 4840 scope.go:117] "RemoveContainer" containerID="c79f4e921ab2d7e8370fe83552c960147cbbd744f03bca3e390813ca3d021aa3" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.588699 4840 scope.go:117] "RemoveContainer" containerID="109628b3a230ce972e436d712479c6f8071e34a983f83d79555ffea902ca2e55" Sep 30 13:57:39 crc kubenswrapper[4840]: E0930 13:57:39.588941 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-kfx69_openshift-ovn-kubernetes(2ff48c28-d076-46e8-a93f-9630989f81e8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.603649 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:39Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.618789 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c2d8050-9531-4413-9d18-14bf3be65acf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34a542c60726f160b423c3e546ff4769a4a7e4a2441b5a09099e4d8c688286c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b5244dfa04a1f1804ef64dc65030d5eaedf679849eec6a9f0731e6cf147f83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7nvs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:39Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.630936 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gp5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1491f559-bc12-4afd-a40c-4eaa40d920a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gp5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:39Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.644524 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49c5addb-0d04-4c3d-b10e-a07d5fec55da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cf852cfa87ef328d172369c26197a9e6a58fe26fa3113469d9275b2e78c30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a08b53dbed5f3a033d2838519733a2a4109ac5a954d9f5fabd0762b2823b090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3f57df6d22c47163751c297aff0ca1dc2bdbf150b9e385e72acdefde92910c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08be626d75cb6bf53342c4ccab7abef9374735cd29d4afb7c7b5b62715db5d22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08be626d75cb6bf53342c4ccab7abef9374735cd29d4afb7c7b5b62715db5d22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:39Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.654441 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.654495 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.654507 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.654526 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.654541 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:39Z","lastTransitionTime":"2025-09-30T13:57:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.659203 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f8cd5bd-4a7f-4a5b-ab00-0cc64de9de94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cba320f443ed90a4e755228b663a1da3096e7ea8e068671d1f6273544b6ef60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c507106f39ae2e8acd4268c4e8f28ffd88a0965e69544d8d8a8830e83eceaf9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d62fbefb67a24fedc5f13652596c76f5b288fec58d79b5209f49c79db139d79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61317d26f40a4361a289779bb97b235525b8dddae383f63abfdcab634b633479\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\"\\\\nI0930 13:56:34.911386 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 13:56:34.920195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 13:56:34.920220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 13:56:34.920240 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 13:56:34.920248 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 13:56:34.927953 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 13:56:34.927977 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927984 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 13:56:34.927992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 13:56:34.927996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 13:56:34.928001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 13:56:34.928159 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0930 13:56:34.930945 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759240579\\\\\\\\\\\\\\\" (2025-09-30 13:56:19 +0000 UTC to 2025-10-30 13:56:20 +0000 UTC (now=2025-09-30 13:56:34.930900435 +0000 UTC))\\\\\\\"\\\\nF0930 13:56:34.930995 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1435e2e88d45e5561ff5863095d0f953ef1478d19d87691ab226cad7ae148f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:39Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.670748 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:39Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.685729 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8d388c29baf0ed22d5c9240f2585e4ca133de265227af6c802a169da6650f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:39Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.698041 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b16edc8f720d4296d7fe50fe407f5983ab975818ec6954d1f0e738e5596624a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:39Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.718580 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84052a424fd6279d70b54902ebdcf430c8aa53b02406b52fc261db9d4d88ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27a94b18137fa991459484406b1c5bd2b2dbb52f527d5f13156ba8e5b8d0146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc4e5d3625aa2d5f94307cedddf72cc48f17f327a044501144e35422d670d555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9549ce85242f1f19b3a52e603caec58bb4c3e2449eb50d8c4525f56dfc93e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a364a13c773fe00ee4e022a286821013d3a681acc16567e66b1c33c7e18ee3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://109628b3a230ce972e436d712479c6f8071e34a983f83d79555ffea902ca2e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c79f4e921ab2d7e8370fe83552c960147cbbd744f03bca3e390813ca3d021aa3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T13:57:09Z\\\",\\\"message\\\":\\\"irewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 13:57:09.078226 6498 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0930 13:57:09.078250 6498 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0930 13:57:09.078275 6498 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0930 13:57:09.078279 6498 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0930 13:57:09.078296 6498 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0930 13:57:09.078313 6498 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0930 13:57:09.078326 6498 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0930 13:57:09.078344 6498 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0930 13:57:09.078531 6498 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0930 13:57:09.078543 6498 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0930 13:57:09.078563 6498 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0930 13:57:09.078571 6498 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 13:57:09.078579 6498 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0930 13:57:09.078585 6498 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0930 13:57:09.078708 6498 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:57:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://109628b3a230ce972e436d712479c6f8071e34a983f83d79555ffea902ca2e55\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T13:57:38Z\\\",\\\"message\\\":\\\"40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0930 13:57:38.886182 6900 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:38Z is after 2025-08-24T17:21:41Z]\\\\nI0930 13:57:38.886183 6900 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0930 13:57:38.886144 6900 model_clien\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:57:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:39Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.732398 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3eb5795ae35b00eba9462e95fdba89c2dcf0442843c32ec05edd710dfe507fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T13:57:29Z\\\",\\\"message\\\":\\\"2025-09-30T13:56:43+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_afea8585-8b55-4126-ad48-e6082b820f13\\\\n2025-09-30T13:56:43+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_afea8585-8b55-4126-ad48-e6082b820f13 to /host/opt/cni/bin/\\\\n2025-09-30T13:56:44Z [verbose] multus-daemon started\\\\n2025-09-30T13:56:44Z [verbose] Readiness Indicator file check\\\\n2025-09-30T13:57:29Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:39Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.756960 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.757019 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.757039 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.757062 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.757080 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:39Z","lastTransitionTime":"2025-09-30T13:57:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.760224 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a4dccdc-e5ba-4cfa-ac4b-e34a27c7368f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15be3c94974cfafcf69a74ef1ff8f6b9da74840805ceafb96aeb903dbf90fd58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://987b6a297870d2898fe2b5df90b15166fb192e10d232fa9bae686ebb7f4180aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a3cd4212aed96f1cad501a5c62065b9c1a289c03afbec269ca193c12434c56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15291f77e6b50679df8feef2f536b5f5745d6f841372770d647831006d8eb0af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d68b17f8a9e41caf54d2f8e2ae775c33d653299e8fe157de0fa4596e1339682e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5a84f046af7784b7ef2afb1b87d108573cd76d20f1b74077166f4dd7594f0b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5a84f046af7784b7ef2afb1b87d108573cd76d20f1b74077166f4dd7594f0b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c2c3a885cd937e387b3ba820ef501fe0c318a277d2353aa6664b3674faf79de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c2c3a885cd937e387b3ba820ef501fe0c318a277d2353aa6664b3674faf79de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f335b20cc833d3278d09697f85b1abd85732aefd80429fb1866eca5565ef98e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f335b20cc833d3278d09697f85b1abd85732aefd80429fb1866eca5565ef98e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:39Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.774248 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:39Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.787032 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:39Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.797977 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:39Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.806728 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pj2zd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4de996b8-bda8-4142-b5e0-04ed9ae1f327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7286e2d685d6573993e2fa000cb9e08d85c6aa820af507327be52ca50fe58c7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pj2zd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:39Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.819398 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:39Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.837633 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:39Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.860073 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.860118 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.860132 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.860152 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.860164 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:39Z","lastTransitionTime":"2025-09-30T13:57:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.962625 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.962693 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.962710 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.962739 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:39 crc kubenswrapper[4840]: I0930 13:57:39.962804 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:39Z","lastTransitionTime":"2025-09-30T13:57:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.065625 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.065684 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.065698 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.065718 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.065731 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:40Z","lastTransitionTime":"2025-09-30T13:57:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.115870 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:57:40 crc kubenswrapper[4840]: E0930 13:57:40.116065 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.168226 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.168279 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.168293 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.168316 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.168329 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:40Z","lastTransitionTime":"2025-09-30T13:57:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.272404 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.272489 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.272505 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.272536 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.272575 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:40Z","lastTransitionTime":"2025-09-30T13:57:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.375815 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.375887 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.375903 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.375928 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.375944 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:40Z","lastTransitionTime":"2025-09-30T13:57:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.480117 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.480186 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.480200 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.480221 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.480236 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:40Z","lastTransitionTime":"2025-09-30T13:57:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.582981 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.583023 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.583032 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.583051 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.583063 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:40Z","lastTransitionTime":"2025-09-30T13:57:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.594969 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kfx69_2ff48c28-d076-46e8-a93f-9630989f81e8/ovnkube-controller/3.log" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.600967 4840 scope.go:117] "RemoveContainer" containerID="109628b3a230ce972e436d712479c6f8071e34a983f83d79555ffea902ca2e55" Sep 30 13:57:40 crc kubenswrapper[4840]: E0930 13:57:40.601281 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-kfx69_openshift-ovn-kubernetes(2ff48c28-d076-46e8-a93f-9630989f81e8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.628728 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84052a424fd6279d70b54902ebdcf430c8aa53b02406b52fc261db9d4d88ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27a94b18137fa991459484406b1c5bd2b2dbb52f527d5f13156ba8e5b8d0146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc4e5d3625aa2d5f94307cedddf72cc48f17f327a044501144e35422d670d555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9549ce85242f1f19b3a52e603caec58bb4c3e2449eb50d8c4525f56dfc93e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a364a13c773fe00ee4e022a286821013d3a681acc16567e66b1c33c7e18ee3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://109628b3a230ce972e436d712479c6f8071e34a983f83d79555ffea902ca2e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://109628b3a230ce972e436d712479c6f8071e34a983f83d79555ffea902ca2e55\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T13:57:38Z\\\",\\\"message\\\":\\\"40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0930 13:57:38.886182 6900 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:38Z is after 2025-08-24T17:21:41Z]\\\\nI0930 13:57:38.886183 6900 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0930 13:57:38.886144 6900 model_clien\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:57:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-kfx69_openshift-ovn-kubernetes(2ff48c28-d076-46e8-a93f-9630989f81e8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:40Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.646542 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3eb5795ae35b00eba9462e95fdba89c2dcf0442843c32ec05edd710dfe507fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T13:57:29Z\\\",\\\"message\\\":\\\"2025-09-30T13:56:43+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_afea8585-8b55-4126-ad48-e6082b820f13\\\\n2025-09-30T13:56:43+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_afea8585-8b55-4126-ad48-e6082b820f13 to /host/opt/cni/bin/\\\\n2025-09-30T13:56:44Z [verbose] multus-daemon started\\\\n2025-09-30T13:56:44Z [verbose] Readiness Indicator file check\\\\n2025-09-30T13:57:29Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:40Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.660167 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pj2zd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4de996b8-bda8-4142-b5e0-04ed9ae1f327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7286e2d685d6573993e2fa000cb9e08d85c6aa820af507327be52ca50fe58c7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pj2zd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:40Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.679789 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a4dccdc-e5ba-4cfa-ac4b-e34a27c7368f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15be3c94974cfafcf69a74ef1ff8f6b9da74840805ceafb96aeb903dbf90fd58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://987b6a297870d2898fe2b5df90b15166fb192e10d232fa9bae686ebb7f4180aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a3cd4212aed96f1cad501a5c62065b9c1a289c03afbec269ca193c12434c56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15291f77e6b50679df8feef2f536b5f5745d6f841372770d647831006d8eb0af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d68b17f8a9e41caf54d2f8e2ae775c33d653299e8fe157de0fa4596e1339682e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5a84f046af7784b7ef2afb1b87d108573cd76d20f1b74077166f4dd7594f0b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5a84f046af7784b7ef2afb1b87d108573cd76d20f1b74077166f4dd7594f0b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c2c3a885cd937e387b3ba820ef501fe0c318a277d2353aa6664b3674faf79de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c2c3a885cd937e387b3ba820ef501fe0c318a277d2353aa6664b3674faf79de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f335b20cc833d3278d09697f85b1abd85732aefd80429fb1866eca5565ef98e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f335b20cc833d3278d09697f85b1abd85732aefd80429fb1866eca5565ef98e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:40Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.686223 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.686279 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.686295 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.686316 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.686332 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:40Z","lastTransitionTime":"2025-09-30T13:57:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.700367 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:40Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.717993 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:40Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.733340 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:40Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.746442 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:40Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.761740 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:40Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.775795 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b16edc8f720d4296d7fe50fe407f5983ab975818ec6954d1f0e738e5596624a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:40Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.786819 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:40Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.793602 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.793632 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.793826 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.793840 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.793849 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:40Z","lastTransitionTime":"2025-09-30T13:57:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.799674 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c2d8050-9531-4413-9d18-14bf3be65acf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34a542c60726f160b423c3e546ff4769a4a7e4a2441b5a09099e4d8c688286c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b5244dfa04a1f1804ef64dc65030d5eaedf679849eec6a9f0731e6cf147f83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7nvs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:40Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.812730 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gp5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1491f559-bc12-4afd-a40c-4eaa40d920a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gp5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:40Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.828851 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49c5addb-0d04-4c3d-b10e-a07d5fec55da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cf852cfa87ef328d172369c26197a9e6a58fe26fa3113469d9275b2e78c30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a08b53dbed5f3a033d2838519733a2a4109ac5a954d9f5fabd0762b2823b090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3f57df6d22c47163751c297aff0ca1dc2bdbf150b9e385e72acdefde92910c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08be626d75cb6bf53342c4ccab7abef9374735cd29d4afb7c7b5b62715db5d22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08be626d75cb6bf53342c4ccab7abef9374735cd29d4afb7c7b5b62715db5d22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:40Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.849665 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f8cd5bd-4a7f-4a5b-ab00-0cc64de9de94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cba320f443ed90a4e755228b663a1da3096e7ea8e068671d1f6273544b6ef60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c507106f39ae2e8acd4268c4e8f28ffd88a0965e69544d8d8a8830e83eceaf9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d62fbefb67a24fedc5f13652596c76f5b288fec58d79b5209f49c79db139d79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61317d26f40a4361a289779bb97b235525b8dddae383f63abfdcab634b633479\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\"\\\\nI0930 13:56:34.911386 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 13:56:34.920195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 13:56:34.920220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 13:56:34.920240 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 13:56:34.920248 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 13:56:34.927953 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 13:56:34.927977 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927984 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 13:56:34.927992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 13:56:34.927996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 13:56:34.928001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 13:56:34.928159 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0930 13:56:34.930945 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759240579\\\\\\\\\\\\\\\" (2025-09-30 13:56:19 +0000 UTC to 2025-10-30 13:56:20 +0000 UTC (now=2025-09-30 13:56:34.930900435 +0000 UTC))\\\\\\\"\\\\nF0930 13:56:34.930995 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1435e2e88d45e5561ff5863095d0f953ef1478d19d87691ab226cad7ae148f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:40Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.859873 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:40Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.874966 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8d388c29baf0ed22d5c9240f2585e4ca133de265227af6c802a169da6650f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:40Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.896190 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.896261 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.896284 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.896314 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.896339 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:40Z","lastTransitionTime":"2025-09-30T13:57:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.999801 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.999861 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:40 crc kubenswrapper[4840]: I0930 13:57:40.999874 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:40.999897 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:40.999911 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:40Z","lastTransitionTime":"2025-09-30T13:57:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.103732 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.103811 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.103824 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.103842 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.103857 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:41Z","lastTransitionTime":"2025-09-30T13:57:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.115893 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.115944 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.116098 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:57:41 crc kubenswrapper[4840]: E0930 13:57:41.116187 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:57:41 crc kubenswrapper[4840]: E0930 13:57:41.116234 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:57:41 crc kubenswrapper[4840]: E0930 13:57:41.116316 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.124540 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.207332 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.207400 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.207417 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.207443 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.207459 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:41Z","lastTransitionTime":"2025-09-30T13:57:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.309917 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.309958 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.309968 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.309982 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.309994 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:41Z","lastTransitionTime":"2025-09-30T13:57:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.413151 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.413225 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.413238 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.413256 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.413269 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:41Z","lastTransitionTime":"2025-09-30T13:57:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.516139 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.516214 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.516236 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.516266 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.516288 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:41Z","lastTransitionTime":"2025-09-30T13:57:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.620176 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.620250 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.620264 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.620289 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.620313 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:41Z","lastTransitionTime":"2025-09-30T13:57:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.723744 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.723827 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.723846 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.723875 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.723893 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:41Z","lastTransitionTime":"2025-09-30T13:57:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.827018 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.827072 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.827083 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.827102 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.827115 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:41Z","lastTransitionTime":"2025-09-30T13:57:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.929672 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.929723 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.929739 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.929760 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:41 crc kubenswrapper[4840]: I0930 13:57:41.929778 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:41Z","lastTransitionTime":"2025-09-30T13:57:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.033006 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.033050 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.033062 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.033082 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.033096 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:42Z","lastTransitionTime":"2025-09-30T13:57:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.115735 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:57:42 crc kubenswrapper[4840]: E0930 13:57:42.116001 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.136783 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.136856 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.136878 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.136909 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.136933 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:42Z","lastTransitionTime":"2025-09-30T13:57:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.137233 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b16edc8f720d4296d7fe50fe407f5983ab975818ec6954d1f0e738e5596624a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.158317 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w988d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da1d702d-bbe7-4e02-aa18-cb6556383674\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4809e039bd5f04b04330e34ae5867e42a0491644ff8338d156beb424eb85cc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p6fs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w988d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.171175 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c2d8050-9531-4413-9d18-14bf3be65acf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34a542c60726f160b423c3e546ff4769a4a7e4a2441b5a09099e4d8c688286c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b5244dfa04a1f1804ef64dc65030d5eaedf679849eec6a9f0731e6cf147f83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wr4j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7nvs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.183967 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4gp5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1491f559-bc12-4afd-a40c-4eaa40d920a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mvf8m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4gp5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.200587 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49c5addb-0d04-4c3d-b10e-a07d5fec55da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cf852cfa87ef328d172369c26197a9e6a58fe26fa3113469d9275b2e78c30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a08b53dbed5f3a033d2838519733a2a4109ac5a954d9f5fabd0762b2823b090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3f57df6d22c47163751c297aff0ca1dc2bdbf150b9e385e72acdefde92910c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08be626d75cb6bf53342c4ccab7abef9374735cd29d4afb7c7b5b62715db5d22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08be626d75cb6bf53342c4ccab7abef9374735cd29d4afb7c7b5b62715db5d22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.215879 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f8cd5bd-4a7f-4a5b-ab00-0cc64de9de94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cba320f443ed90a4e755228b663a1da3096e7ea8e068671d1f6273544b6ef60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c507106f39ae2e8acd4268c4e8f28ffd88a0965e69544d8d8a8830e83eceaf9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d62fbefb67a24fedc5f13652596c76f5b288fec58d79b5209f49c79db139d79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61317d26f40a4361a289779bb97b235525b8dddae383f63abfdcab634b633479\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbb685de55486486b0da9c750722e9ec27458f2582787e37080a30edcde807eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T13:56:34Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\"\\\\nI0930 13:56:34.911386 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 13:56:34.920195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 13:56:34.920220 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 13:56:34.920240 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 13:56:34.920248 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 13:56:34.927953 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 13:56:34.927977 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927984 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 13:56:34.927989 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 13:56:34.927992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 13:56:34.927996 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 13:56:34.928001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 13:56:34.928159 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0930 13:56:34.930945 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3607038965/tls.crt::/tmp/serving-cert-3607038965/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759240579\\\\\\\\\\\\\\\" (2025-09-30 13:56:19 +0000 UTC to 2025-10-30 13:56:20 +0000 UTC (now=2025-09-30 13:56:34.930900435 +0000 UTC))\\\\\\\"\\\\nF0930 13:56:34.930995 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1435e2e88d45e5561ff5863095d0f953ef1478d19d87691ab226cad7ae148f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61952967f0dea5c74ab8e2a27212add5bd721825d28b97960598869b1ccb21f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.230184 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e8b890-7f20-4a36-8e03-898620cf599a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91d4d40fbad439e23edc61624db2fcfa7561962bedf32c696963ba80b60fbb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv2rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-747gk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.239885 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.239925 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.239935 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.239951 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.239962 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:42Z","lastTransitionTime":"2025-09-30T13:57:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.246718 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b264ec9-951e-4928-b43c-5c045b7681f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8d388c29baf0ed22d5c9240f2585e4ca133de265227af6c802a169da6650f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://642e9927cb0eb39a38220ce6fc84e2bea56415b48ff23eb47ad308d30d3c5be8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d36ea40ee166bde8c2a817d260fb6acfeac81691c44a1f62575cde323f207ddb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1898503ba9474b791e321ef0c66eb631aa4227b8d260a94e689f4054a3c79228\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://210a38598d13c79f1452196762e4e31cfaf7d316f2cb770b15dbd07fb821e9b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3632f17724b0d2d374e5146fcf0e86bacddc43098976c39c63a52269cd7bef5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6fcac92fcc63695983cfccc06f18c728b9e751c44fa6aa12830dc48c283834e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6q8j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hmwzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.268718 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ff48c28-d076-46e8-a93f-9630989f81e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84052a424fd6279d70b54902ebdcf430c8aa53b02406b52fc261db9d4d88ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f27a94b18137fa991459484406b1c5bd2b2dbb52f527d5f13156ba8e5b8d0146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc4e5d3625aa2d5f94307cedddf72cc48f17f327a044501144e35422d670d555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9549ce85242f1f19b3a52e603caec58bb4c3e2449eb50d8c4525f56dfc93e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a364a13c773fe00ee4e022a286821013d3a681acc16567e66b1c33c7e18ee3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://109628b3a230ce972e436d712479c6f8071e34a983f83d79555ffea902ca2e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://109628b3a230ce972e436d712479c6f8071e34a983f83d79555ffea902ca2e55\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T13:57:38Z\\\",\\\"message\\\":\\\"40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0930 13:57:38.886182 6900 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:38Z is after 2025-08-24T17:21:41Z]\\\\nI0930 13:57:38.886183 6900 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0930 13:57:38.886144 6900 model_clien\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:57:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-kfx69_openshift-ovn-kubernetes(2ff48c28-d076-46e8-a93f-9630989f81e8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tbvzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kfx69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.284543 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwvl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe233ee-1ea7-433e-a53a-e4a668f739ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3eb5795ae35b00eba9462e95fdba89c2dcf0442843c32ec05edd710dfe507fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T13:57:29Z\\\",\\\"message\\\":\\\"2025-09-30T13:56:43+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_afea8585-8b55-4126-ad48-e6082b820f13\\\\n2025-09-30T13:56:43+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_afea8585-8b55-4126-ad48-e6082b820f13 to /host/opt/cni/bin/\\\\n2025-09-30T13:56:44Z [verbose] multus-daemon started\\\\n2025-09-30T13:56:44Z [verbose] Readiness Indicator file check\\\\n2025-09-30T13:57:29Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:57:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndw2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwvl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.296596 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pj2zd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4de996b8-bda8-4142-b5e0-04ed9ae1f327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7286e2d685d6573993e2fa000cb9e08d85c6aa820af507327be52ca50fe58c7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pj2zd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.316376 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a4dccdc-e5ba-4cfa-ac4b-e34a27c7368f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15be3c94974cfafcf69a74ef1ff8f6b9da74840805ceafb96aeb903dbf90fd58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://987b6a297870d2898fe2b5df90b15166fb192e10d232fa9bae686ebb7f4180aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a3cd4212aed96f1cad501a5c62065b9c1a289c03afbec269ca193c12434c56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15291f77e6b50679df8feef2f536b5f5745d6f841372770d647831006d8eb0af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d68b17f8a9e41caf54d2f8e2ae775c33d653299e8fe157de0fa4596e1339682e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5a84f046af7784b7ef2afb1b87d108573cd76d20f1b74077166f4dd7594f0b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5a84f046af7784b7ef2afb1b87d108573cd76d20f1b74077166f4dd7594f0b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c2c3a885cd937e387b3ba820ef501fe0c318a277d2353aa6664b3674faf79de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c2c3a885cd937e387b3ba820ef501fe0c318a277d2353aa6664b3674faf79de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f335b20cc833d3278d09697f85b1abd85732aefd80429fb1866eca5565ef98e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f335b20cc833d3278d09697f85b1abd85732aefd80429fb1866eca5565ef98e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.332446 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5524de25aacb3bef68eb40a562bbef57583518c79f8a09f2e90f8397095f14ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.342444 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.342507 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.342520 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.342570 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.342591 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:42Z","lastTransitionTime":"2025-09-30T13:57:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.349180 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.367708 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.380528 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e34a533e-045f-49e1-9a7e-ae836d05ca09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c8399ebe59b7663b0af7b042b0b4e67fcdf61efe10c803f50b1184fce406456\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://567d4e94a770ba3318053366878e777f03b826d75569ada21be35f83e4f21740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://567d4e94a770ba3318053366878e777f03b826d75569ada21be35f83e4f21740\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T13:56:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T13:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T13:56:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.406077 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.424985 4840 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T13:56:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0c933b7bf17fd711b8cb1f225f9f006666606ca53f08b701c8acfe465ae5c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4813c494c2f3af7d7d449005ad3759248e2372f55fb00d3acc246db2beee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T13:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:42Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.445239 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.445310 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.445336 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.445368 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.445391 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:42Z","lastTransitionTime":"2025-09-30T13:57:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.548099 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.548167 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.548178 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.548199 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.548216 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:42Z","lastTransitionTime":"2025-09-30T13:57:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.650504 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.650545 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.650567 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.650582 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.650591 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:42Z","lastTransitionTime":"2025-09-30T13:57:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.753486 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.753528 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.753540 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.753572 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.753584 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:42Z","lastTransitionTime":"2025-09-30T13:57:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.856287 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.856324 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.856344 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.856364 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.856375 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:42Z","lastTransitionTime":"2025-09-30T13:57:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.959670 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.959751 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.959774 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.959804 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.959828 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:42Z","lastTransitionTime":"2025-09-30T13:57:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.982461 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:57:42 crc kubenswrapper[4840]: E0930 13:57:42.982720 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:46.982680685 +0000 UTC m=+155.611767148 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.982791 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.982848 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.982894 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:57:42 crc kubenswrapper[4840]: I0930 13:57:42.982940 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:57:42 crc kubenswrapper[4840]: E0930 13:57:42.983023 4840 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 13:57:42 crc kubenswrapper[4840]: E0930 13:57:42.983073 4840 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 13:57:42 crc kubenswrapper[4840]: E0930 13:57:42.983104 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 13:58:46.983084886 +0000 UTC m=+155.612171349 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 13:57:42 crc kubenswrapper[4840]: E0930 13:57:42.983137 4840 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 13:57:42 crc kubenswrapper[4840]: E0930 13:57:42.983173 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 13:58:46.983145918 +0000 UTC m=+155.612232391 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 13:57:42 crc kubenswrapper[4840]: E0930 13:57:42.983181 4840 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 13:57:42 crc kubenswrapper[4840]: E0930 13:57:42.983202 4840 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 13:57:42 crc kubenswrapper[4840]: E0930 13:57:42.983226 4840 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 13:57:42 crc kubenswrapper[4840]: E0930 13:57:42.983260 4840 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 13:57:42 crc kubenswrapper[4840]: E0930 13:57:42.983269 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-30 13:58:46.983250241 +0000 UTC m=+155.612336744 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 13:57:42 crc kubenswrapper[4840]: E0930 13:57:42.983273 4840 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 13:57:42 crc kubenswrapper[4840]: E0930 13:57:42.983403 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-30 13:58:46.983356663 +0000 UTC m=+155.612443086 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.063434 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.063487 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.063499 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.063516 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.063530 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:43Z","lastTransitionTime":"2025-09-30T13:57:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.115757 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.115913 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.115968 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:57:43 crc kubenswrapper[4840]: E0930 13:57:43.116138 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:57:43 crc kubenswrapper[4840]: E0930 13:57:43.116391 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:57:43 crc kubenswrapper[4840]: E0930 13:57:43.116540 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.165752 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.165800 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.165813 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.165833 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.165845 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:43Z","lastTransitionTime":"2025-09-30T13:57:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.269149 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.269218 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.269236 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.269264 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.269280 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:43Z","lastTransitionTime":"2025-09-30T13:57:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.372179 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.372214 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.372225 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.372239 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.372250 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:43Z","lastTransitionTime":"2025-09-30T13:57:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.475625 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.475688 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.475706 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.475757 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.475778 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:43Z","lastTransitionTime":"2025-09-30T13:57:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.579175 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.579331 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.579359 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.579389 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.579434 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:43Z","lastTransitionTime":"2025-09-30T13:57:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.682233 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.682271 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.682280 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.682294 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.682305 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:43Z","lastTransitionTime":"2025-09-30T13:57:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.784508 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.784585 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.784602 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.784621 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.784633 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:43Z","lastTransitionTime":"2025-09-30T13:57:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.886870 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.886916 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.886927 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.886944 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.886956 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:43Z","lastTransitionTime":"2025-09-30T13:57:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.989016 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.989050 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.989058 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.989073 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:43 crc kubenswrapper[4840]: I0930 13:57:43.989082 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:43Z","lastTransitionTime":"2025-09-30T13:57:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.092011 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.092092 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.092114 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.092142 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.092163 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:44Z","lastTransitionTime":"2025-09-30T13:57:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.115740 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:57:44 crc kubenswrapper[4840]: E0930 13:57:44.115874 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.194333 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.194372 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.194380 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.194399 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.194410 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:44Z","lastTransitionTime":"2025-09-30T13:57:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.297357 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.297415 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.297425 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.297438 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.297447 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:44Z","lastTransitionTime":"2025-09-30T13:57:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.400471 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.400524 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.400543 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.400608 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.400630 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:44Z","lastTransitionTime":"2025-09-30T13:57:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.502965 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.503000 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.503011 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.503023 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.503033 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:44Z","lastTransitionTime":"2025-09-30T13:57:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.606209 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.606247 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.606257 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.606275 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.606286 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:44Z","lastTransitionTime":"2025-09-30T13:57:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.708743 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.708790 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.708800 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.708816 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.708827 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:44Z","lastTransitionTime":"2025-09-30T13:57:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.811340 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.811404 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.811421 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.811447 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.811464 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:44Z","lastTransitionTime":"2025-09-30T13:57:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.914868 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.914933 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.914950 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.914973 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:44 crc kubenswrapper[4840]: I0930 13:57:44.914991 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:44Z","lastTransitionTime":"2025-09-30T13:57:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.017388 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.017428 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.017436 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.017451 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.017462 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:45Z","lastTransitionTime":"2025-09-30T13:57:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.115992 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.116055 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:57:45 crc kubenswrapper[4840]: E0930 13:57:45.116347 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.116055 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:57:45 crc kubenswrapper[4840]: E0930 13:57:45.116455 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:57:45 crc kubenswrapper[4840]: E0930 13:57:45.116575 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.119670 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.119705 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.119716 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.119733 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.119745 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:45Z","lastTransitionTime":"2025-09-30T13:57:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.223092 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.223171 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.223188 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.223212 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.223227 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:45Z","lastTransitionTime":"2025-09-30T13:57:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.325974 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.326012 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.326020 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.326038 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.326053 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:45Z","lastTransitionTime":"2025-09-30T13:57:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.429269 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.429308 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.429317 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.429333 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.429343 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:45Z","lastTransitionTime":"2025-09-30T13:57:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.533444 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.533523 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.533541 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.533593 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.533611 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:45Z","lastTransitionTime":"2025-09-30T13:57:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.636724 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.636847 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.636873 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.636903 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.636929 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:45Z","lastTransitionTime":"2025-09-30T13:57:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.740809 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.740877 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.740895 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.740919 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.740937 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:45Z","lastTransitionTime":"2025-09-30T13:57:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.843678 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.843765 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.843789 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.843836 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.843860 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:45Z","lastTransitionTime":"2025-09-30T13:57:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.946876 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.946938 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.946962 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.946991 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:45 crc kubenswrapper[4840]: I0930 13:57:45.947013 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:45Z","lastTransitionTime":"2025-09-30T13:57:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.050987 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.051041 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.051054 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.051076 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.051089 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:46Z","lastTransitionTime":"2025-09-30T13:57:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.115978 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:57:46 crc kubenswrapper[4840]: E0930 13:57:46.116152 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.153136 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.153177 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.153187 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.153202 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.153213 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:46Z","lastTransitionTime":"2025-09-30T13:57:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.255443 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.255482 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.255495 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.255511 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.255522 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:46Z","lastTransitionTime":"2025-09-30T13:57:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.358526 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.358630 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.358653 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.358680 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.358703 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:46Z","lastTransitionTime":"2025-09-30T13:57:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.462017 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.462092 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.462111 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.462134 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.462150 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:46Z","lastTransitionTime":"2025-09-30T13:57:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.565645 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.565691 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.565703 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.565725 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.565747 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:46Z","lastTransitionTime":"2025-09-30T13:57:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.669470 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.669523 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.669533 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.669572 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.669586 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:46Z","lastTransitionTime":"2025-09-30T13:57:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.772816 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.772895 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.772910 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.772944 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.772962 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:46Z","lastTransitionTime":"2025-09-30T13:57:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.876411 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.876466 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.876477 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.876501 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.876515 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:46Z","lastTransitionTime":"2025-09-30T13:57:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.980152 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.980212 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.980224 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.980244 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:46 crc kubenswrapper[4840]: I0930 13:57:46.980256 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:46Z","lastTransitionTime":"2025-09-30T13:57:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.084289 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.084360 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.084372 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.084386 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.084400 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:47Z","lastTransitionTime":"2025-09-30T13:57:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.116058 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.116114 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.116142 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:57:47 crc kubenswrapper[4840]: E0930 13:57:47.116218 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:57:47 crc kubenswrapper[4840]: E0930 13:57:47.116315 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:57:47 crc kubenswrapper[4840]: E0930 13:57:47.116413 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.136710 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.136778 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.136793 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.136811 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.136822 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:47Z","lastTransitionTime":"2025-09-30T13:57:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:47 crc kubenswrapper[4840]: E0930 13:57:47.150525 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5\\\",\\\"systemUUID\\\":\\\"f41e0351-4eda-411a-9a99-43a1952f3d34\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:47Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.154290 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.154322 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.154335 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.154351 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.154363 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:47Z","lastTransitionTime":"2025-09-30T13:57:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:47 crc kubenswrapper[4840]: E0930 13:57:47.168444 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5\\\",\\\"systemUUID\\\":\\\"f41e0351-4eda-411a-9a99-43a1952f3d34\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:47Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.172916 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.172978 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.172988 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.173009 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.173019 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:47Z","lastTransitionTime":"2025-09-30T13:57:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:47 crc kubenswrapper[4840]: E0930 13:57:47.185762 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5\\\",\\\"systemUUID\\\":\\\"f41e0351-4eda-411a-9a99-43a1952f3d34\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:47Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.189503 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.189538 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.189560 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.189576 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.189585 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:47Z","lastTransitionTime":"2025-09-30T13:57:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:47 crc kubenswrapper[4840]: E0930 13:57:47.202203 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5\\\",\\\"systemUUID\\\":\\\"f41e0351-4eda-411a-9a99-43a1952f3d34\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:47Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.205336 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.205377 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.205388 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.205405 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.205415 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:47Z","lastTransitionTime":"2025-09-30T13:57:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:47 crc kubenswrapper[4840]: E0930 13:57:47.217438 4840 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T13:57:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"70ad9f1f-eb7d-4df2-8e0e-21d7c9c559b5\\\",\\\"systemUUID\\\":\\\"f41e0351-4eda-411a-9a99-43a1952f3d34\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T13:57:47Z is after 2025-08-24T17:21:41Z" Sep 30 13:57:47 crc kubenswrapper[4840]: E0930 13:57:47.217604 4840 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.219128 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.219177 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.219197 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.219223 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.219240 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:47Z","lastTransitionTime":"2025-09-30T13:57:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.321195 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.321299 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.321319 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.321336 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.321346 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:47Z","lastTransitionTime":"2025-09-30T13:57:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.424348 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.424380 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.424389 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.424406 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.424417 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:47Z","lastTransitionTime":"2025-09-30T13:57:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.526884 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.526967 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.526991 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.527022 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.527045 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:47Z","lastTransitionTime":"2025-09-30T13:57:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.629757 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.629842 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.629870 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.629901 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.629926 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:47Z","lastTransitionTime":"2025-09-30T13:57:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.732625 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.732670 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.732682 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.732699 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.732711 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:47Z","lastTransitionTime":"2025-09-30T13:57:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.836221 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.836267 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.836286 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.836307 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.836322 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:47Z","lastTransitionTime":"2025-09-30T13:57:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.938886 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.939190 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.939285 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.939377 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:47 crc kubenswrapper[4840]: I0930 13:57:47.939454 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:47Z","lastTransitionTime":"2025-09-30T13:57:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.042494 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.042528 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.042538 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.042586 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.042601 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:48Z","lastTransitionTime":"2025-09-30T13:57:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.115662 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:57:48 crc kubenswrapper[4840]: E0930 13:57:48.115889 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.145149 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.145204 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.145216 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.145234 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.145247 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:48Z","lastTransitionTime":"2025-09-30T13:57:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.248294 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.248375 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.248394 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.248424 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.248446 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:48Z","lastTransitionTime":"2025-09-30T13:57:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.351595 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.351641 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.351652 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.351667 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.351679 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:48Z","lastTransitionTime":"2025-09-30T13:57:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.455049 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.455133 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.455156 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.455187 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.455212 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:48Z","lastTransitionTime":"2025-09-30T13:57:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.558539 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.558643 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.558665 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.558702 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.558725 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:48Z","lastTransitionTime":"2025-09-30T13:57:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.661178 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.661239 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.661255 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.661278 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.661300 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:48Z","lastTransitionTime":"2025-09-30T13:57:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.765148 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.765227 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.765256 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.765283 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.765303 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:48Z","lastTransitionTime":"2025-09-30T13:57:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.868294 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.868338 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.868346 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.868360 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.868370 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:48Z","lastTransitionTime":"2025-09-30T13:57:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.972146 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.972226 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.972249 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.972278 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:48 crc kubenswrapper[4840]: I0930 13:57:48.972300 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:48Z","lastTransitionTime":"2025-09-30T13:57:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.075852 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.075920 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.075937 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.076142 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.076159 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:49Z","lastTransitionTime":"2025-09-30T13:57:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.115476 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:57:49 crc kubenswrapper[4840]: E0930 13:57:49.115743 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.116011 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.116016 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:57:49 crc kubenswrapper[4840]: E0930 13:57:49.116284 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:57:49 crc kubenswrapper[4840]: E0930 13:57:49.116602 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.127125 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.179332 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.179392 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.179409 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.179432 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.179454 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:49Z","lastTransitionTime":"2025-09-30T13:57:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.282127 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.282185 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.282205 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.282228 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.282252 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:49Z","lastTransitionTime":"2025-09-30T13:57:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.385385 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.385431 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.385442 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.385458 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.385470 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:49Z","lastTransitionTime":"2025-09-30T13:57:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.488127 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.488175 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.488187 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.488202 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.488212 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:49Z","lastTransitionTime":"2025-09-30T13:57:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.590101 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.590134 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.590145 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.590160 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.590169 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:49Z","lastTransitionTime":"2025-09-30T13:57:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.692756 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.692802 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.692813 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.692832 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.692843 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:49Z","lastTransitionTime":"2025-09-30T13:57:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.794862 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.794910 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.794920 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.794937 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.794946 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:49Z","lastTransitionTime":"2025-09-30T13:57:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.897773 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.897817 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.897827 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.897844 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:49 crc kubenswrapper[4840]: I0930 13:57:49.897854 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:49Z","lastTransitionTime":"2025-09-30T13:57:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.001539 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.001625 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.001638 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.001654 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.001666 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:50Z","lastTransitionTime":"2025-09-30T13:57:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.105301 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.105379 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.105396 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.105422 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.105442 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:50Z","lastTransitionTime":"2025-09-30T13:57:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.115962 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:57:50 crc kubenswrapper[4840]: E0930 13:57:50.116176 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.207933 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.208004 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.208021 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.208045 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.208063 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:50Z","lastTransitionTime":"2025-09-30T13:57:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.311174 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.311216 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.311230 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.311252 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.311264 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:50Z","lastTransitionTime":"2025-09-30T13:57:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.414289 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.414366 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.414388 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.414415 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.414435 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:50Z","lastTransitionTime":"2025-09-30T13:57:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.517106 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.517163 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.517181 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.517204 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.517222 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:50Z","lastTransitionTime":"2025-09-30T13:57:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.620829 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.620899 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.620922 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.620951 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.620972 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:50Z","lastTransitionTime":"2025-09-30T13:57:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.724423 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.724498 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.724523 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.724590 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.724621 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:50Z","lastTransitionTime":"2025-09-30T13:57:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.828217 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.828284 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.828301 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.828327 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.828343 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:50Z","lastTransitionTime":"2025-09-30T13:57:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.931259 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.931341 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.931361 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.931388 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:50 crc kubenswrapper[4840]: I0930 13:57:50.931407 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:50Z","lastTransitionTime":"2025-09-30T13:57:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.034493 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.034900 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.034929 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.034954 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.034972 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:51Z","lastTransitionTime":"2025-09-30T13:57:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.115950 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.116016 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.115990 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:57:51 crc kubenswrapper[4840]: E0930 13:57:51.116198 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:57:51 crc kubenswrapper[4840]: E0930 13:57:51.116332 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:57:51 crc kubenswrapper[4840]: E0930 13:57:51.116407 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.139118 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.139203 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.139240 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.139273 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.139297 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:51Z","lastTransitionTime":"2025-09-30T13:57:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.242797 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.242860 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.242877 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.242900 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.242918 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:51Z","lastTransitionTime":"2025-09-30T13:57:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.345318 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.345412 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.345429 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.345454 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.345471 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:51Z","lastTransitionTime":"2025-09-30T13:57:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.448604 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.448685 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.448709 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.448755 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.448779 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:51Z","lastTransitionTime":"2025-09-30T13:57:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.552096 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.552163 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.552174 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.552188 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.552198 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:51Z","lastTransitionTime":"2025-09-30T13:57:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.655191 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.655249 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.655265 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.655291 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.655312 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:51Z","lastTransitionTime":"2025-09-30T13:57:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.757994 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.758056 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.758074 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.758099 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.758116 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:51Z","lastTransitionTime":"2025-09-30T13:57:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.860315 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.860354 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.860366 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.860383 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.860397 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:51Z","lastTransitionTime":"2025-09-30T13:57:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.963608 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.963679 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.963697 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.963721 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:51 crc kubenswrapper[4840]: I0930 13:57:51.963738 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:51Z","lastTransitionTime":"2025-09-30T13:57:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.066234 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.066289 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.066317 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.066340 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.066355 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:52Z","lastTransitionTime":"2025-09-30T13:57:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.116405 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:57:52 crc kubenswrapper[4840]: E0930 13:57:52.116526 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.161038 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=11.161010411 podStartE2EDuration="11.161010411s" podCreationTimestamp="2025-09-30 13:57:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:57:52.160921459 +0000 UTC m=+100.790007942" watchObservedRunningTime="2025-09-30 13:57:52.161010411 +0000 UTC m=+100.790096874" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.169464 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.169511 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.169529 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.169571 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.169586 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:52Z","lastTransitionTime":"2025-09-30T13:57:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.198011 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podStartSLOduration=78.19799036 podStartE2EDuration="1m18.19799036s" podCreationTimestamp="2025-09-30 13:56:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:57:52.1979845 +0000 UTC m=+100.827070923" watchObservedRunningTime="2025-09-30 13:57:52.19799036 +0000 UTC m=+100.827076783" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.215798 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-hmwzh" podStartSLOduration=78.215781141 podStartE2EDuration="1m18.215781141s" podCreationTimestamp="2025-09-30 13:56:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:57:52.21574083 +0000 UTC m=+100.844827253" watchObservedRunningTime="2025-09-30 13:57:52.215781141 +0000 UTC m=+100.844867564" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.246696 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-w988d" podStartSLOduration=78.246674369 podStartE2EDuration="1m18.246674369s" podCreationTimestamp="2025-09-30 13:56:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:57:52.246218127 +0000 UTC m=+100.875304540" watchObservedRunningTime="2025-09-30 13:57:52.246674369 +0000 UTC m=+100.875760802" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.269860 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7nvs5" podStartSLOduration=77.269842913 podStartE2EDuration="1m17.269842913s" podCreationTimestamp="2025-09-30 13:56:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:57:52.269781341 +0000 UTC m=+100.898867774" watchObservedRunningTime="2025-09-30 13:57:52.269842913 +0000 UTC m=+100.898929336" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.271366 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.271391 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.271400 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.271412 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.271422 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:52Z","lastTransitionTime":"2025-09-30T13:57:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.299261 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=45.299244811 podStartE2EDuration="45.299244811s" podCreationTimestamp="2025-09-30 13:57:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:57:52.298294356 +0000 UTC m=+100.927380779" watchObservedRunningTime="2025-09-30 13:57:52.299244811 +0000 UTC m=+100.928331244" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.312756 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=71.312736828 podStartE2EDuration="1m11.312736828s" podCreationTimestamp="2025-09-30 13:56:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:57:52.312625655 +0000 UTC m=+100.941712078" watchObservedRunningTime="2025-09-30 13:57:52.312736828 +0000 UTC m=+100.941823251" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.324560 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-bwvl2" podStartSLOduration=78.32452795 podStartE2EDuration="1m18.32452795s" podCreationTimestamp="2025-09-30 13:56:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:57:52.324247763 +0000 UTC m=+100.953334186" watchObservedRunningTime="2025-09-30 13:57:52.32452795 +0000 UTC m=+100.953614373" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.341254 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=3.341237603 podStartE2EDuration="3.341237603s" podCreationTimestamp="2025-09-30 13:57:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:57:52.341038227 +0000 UTC m=+100.970124650" watchObservedRunningTime="2025-09-30 13:57:52.341237603 +0000 UTC m=+100.970324026" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.373091 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.373136 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.373144 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.373160 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.373171 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:52Z","lastTransitionTime":"2025-09-30T13:57:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.420194 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-pj2zd" podStartSLOduration=78.420174033 podStartE2EDuration="1m18.420174033s" podCreationTimestamp="2025-09-30 13:56:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:57:52.398994542 +0000 UTC m=+101.028080965" watchObservedRunningTime="2025-09-30 13:57:52.420174033 +0000 UTC m=+101.049260456" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.420409 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=22.420401639 podStartE2EDuration="22.420401639s" podCreationTimestamp="2025-09-30 13:57:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:57:52.418874698 +0000 UTC m=+101.047961121" watchObservedRunningTime="2025-09-30 13:57:52.420401639 +0000 UTC m=+101.049488062" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.475773 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.475801 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.475809 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.475847 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.475862 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:52Z","lastTransitionTime":"2025-09-30T13:57:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.578382 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.578426 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.578438 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.578453 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.578465 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:52Z","lastTransitionTime":"2025-09-30T13:57:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.681739 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.682177 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.682408 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.682651 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.682856 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:52Z","lastTransitionTime":"2025-09-30T13:57:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.785601 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.785897 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.785991 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.786203 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.786312 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:52Z","lastTransitionTime":"2025-09-30T13:57:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.889975 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.890251 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.890398 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.890530 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.890713 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:52Z","lastTransitionTime":"2025-09-30T13:57:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.994613 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.994662 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.994673 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.994691 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:52 crc kubenswrapper[4840]: I0930 13:57:52.994706 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:52Z","lastTransitionTime":"2025-09-30T13:57:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.096958 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.097006 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.097017 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.097034 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.097049 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:53Z","lastTransitionTime":"2025-09-30T13:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.115363 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.115424 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:57:53 crc kubenswrapper[4840]: E0930 13:57:53.115488 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.115608 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:57:53 crc kubenswrapper[4840]: E0930 13:57:53.115783 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:57:53 crc kubenswrapper[4840]: E0930 13:57:53.115874 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.116925 4840 scope.go:117] "RemoveContainer" containerID="109628b3a230ce972e436d712479c6f8071e34a983f83d79555ffea902ca2e55" Sep 30 13:57:53 crc kubenswrapper[4840]: E0930 13:57:53.117178 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-kfx69_openshift-ovn-kubernetes(2ff48c28-d076-46e8-a93f-9630989f81e8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.199285 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.199325 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.199333 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.199349 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.199361 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:53Z","lastTransitionTime":"2025-09-30T13:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.301504 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.301548 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.301579 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.301614 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.301627 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:53Z","lastTransitionTime":"2025-09-30T13:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.404179 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.404259 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.404294 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.404323 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.404345 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:53Z","lastTransitionTime":"2025-09-30T13:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.506905 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.506944 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.506956 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.506973 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.506985 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:53Z","lastTransitionTime":"2025-09-30T13:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.609610 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.609660 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.609671 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.609687 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.609697 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:53Z","lastTransitionTime":"2025-09-30T13:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.713051 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.713114 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.713125 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.713142 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.713156 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:53Z","lastTransitionTime":"2025-09-30T13:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.804432 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1491f559-bc12-4afd-a40c-4eaa40d920a8-metrics-certs\") pod \"network-metrics-daemon-4gp5x\" (UID: \"1491f559-bc12-4afd-a40c-4eaa40d920a8\") " pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:57:53 crc kubenswrapper[4840]: E0930 13:57:53.804605 4840 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 13:57:53 crc kubenswrapper[4840]: E0930 13:57:53.804675 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1491f559-bc12-4afd-a40c-4eaa40d920a8-metrics-certs podName:1491f559-bc12-4afd-a40c-4eaa40d920a8 nodeName:}" failed. No retries permitted until 2025-09-30 13:58:57.804656555 +0000 UTC m=+166.433742978 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1491f559-bc12-4afd-a40c-4eaa40d920a8-metrics-certs") pod "network-metrics-daemon-4gp5x" (UID: "1491f559-bc12-4afd-a40c-4eaa40d920a8") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.816202 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.816270 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.816288 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.816312 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.816335 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:53Z","lastTransitionTime":"2025-09-30T13:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.918997 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.919047 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.919060 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.919078 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:53 crc kubenswrapper[4840]: I0930 13:57:53.919090 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:53Z","lastTransitionTime":"2025-09-30T13:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.020995 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.021046 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.021065 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.021086 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.021101 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:54Z","lastTransitionTime":"2025-09-30T13:57:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.115533 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:57:54 crc kubenswrapper[4840]: E0930 13:57:54.115752 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.123574 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.123622 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.123634 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.123650 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.123663 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:54Z","lastTransitionTime":"2025-09-30T13:57:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.227139 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.227189 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.227205 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.227230 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.227251 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:54Z","lastTransitionTime":"2025-09-30T13:57:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.330766 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.330828 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.330841 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.330870 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.330893 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:54Z","lastTransitionTime":"2025-09-30T13:57:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.433975 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.434043 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.434064 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.434095 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.434121 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:54Z","lastTransitionTime":"2025-09-30T13:57:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.537015 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.537084 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.537106 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.537135 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.537159 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:54Z","lastTransitionTime":"2025-09-30T13:57:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.639721 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.639770 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.639779 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.639794 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.639806 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:54Z","lastTransitionTime":"2025-09-30T13:57:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.742969 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.743032 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.743045 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.743063 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.743093 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:54Z","lastTransitionTime":"2025-09-30T13:57:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.845468 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.845877 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.845903 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.845925 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.845942 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:54Z","lastTransitionTime":"2025-09-30T13:57:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.948194 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.948256 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.948273 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.948298 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:54 crc kubenswrapper[4840]: I0930 13:57:54.948315 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:54Z","lastTransitionTime":"2025-09-30T13:57:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.051109 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.051144 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.051152 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.051169 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.051179 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:55Z","lastTransitionTime":"2025-09-30T13:57:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.115886 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.115947 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.115902 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:57:55 crc kubenswrapper[4840]: E0930 13:57:55.116070 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:57:55 crc kubenswrapper[4840]: E0930 13:57:55.116175 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:57:55 crc kubenswrapper[4840]: E0930 13:57:55.116365 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.153688 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.153735 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.153746 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.153760 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.153771 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:55Z","lastTransitionTime":"2025-09-30T13:57:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.255936 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.255990 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.256003 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.256021 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.256033 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:55Z","lastTransitionTime":"2025-09-30T13:57:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.358368 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.358411 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.358424 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.358439 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.358450 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:55Z","lastTransitionTime":"2025-09-30T13:57:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.461808 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.461865 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.461874 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.461888 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.461900 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:55Z","lastTransitionTime":"2025-09-30T13:57:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.564475 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.564516 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.564536 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.564574 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.564589 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:55Z","lastTransitionTime":"2025-09-30T13:57:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.667902 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.667942 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.667952 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.667968 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.667978 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:55Z","lastTransitionTime":"2025-09-30T13:57:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.771437 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.771475 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.771490 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.771508 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.771523 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:55Z","lastTransitionTime":"2025-09-30T13:57:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.875487 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.875542 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.875570 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.875585 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.875596 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:55Z","lastTransitionTime":"2025-09-30T13:57:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.978783 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.978824 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.978836 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.978852 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:55 crc kubenswrapper[4840]: I0930 13:57:55.978864 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:55Z","lastTransitionTime":"2025-09-30T13:57:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.081465 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.081506 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.081519 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.081534 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.081568 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:56Z","lastTransitionTime":"2025-09-30T13:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.115704 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:57:56 crc kubenswrapper[4840]: E0930 13:57:56.115967 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.184104 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.184140 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.184149 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.184163 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.184172 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:56Z","lastTransitionTime":"2025-09-30T13:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.286879 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.286952 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.286975 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.286998 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.287017 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:56Z","lastTransitionTime":"2025-09-30T13:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.390020 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.390087 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.390104 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.390131 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.390148 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:56Z","lastTransitionTime":"2025-09-30T13:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.491986 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.492032 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.492045 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.492062 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.492076 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:56Z","lastTransitionTime":"2025-09-30T13:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.595424 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.595470 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.595479 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.595495 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.595506 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:56Z","lastTransitionTime":"2025-09-30T13:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.698188 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.698233 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.698248 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.698266 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.698279 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:56Z","lastTransitionTime":"2025-09-30T13:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.802142 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.802214 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.802238 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.802271 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.802298 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:56Z","lastTransitionTime":"2025-09-30T13:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.905334 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.905431 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.905458 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.905489 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:56 crc kubenswrapper[4840]: I0930 13:57:56.905514 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:56Z","lastTransitionTime":"2025-09-30T13:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.007795 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.007827 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.007835 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.007849 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.007858 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:57Z","lastTransitionTime":"2025-09-30T13:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.112782 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.112845 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.112858 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.112874 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.112885 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:57Z","lastTransitionTime":"2025-09-30T13:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.116211 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.116234 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.116220 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:57:57 crc kubenswrapper[4840]: E0930 13:57:57.116374 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:57:57 crc kubenswrapper[4840]: E0930 13:57:57.116456 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:57:57 crc kubenswrapper[4840]: E0930 13:57:57.116643 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.215050 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.215128 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.215145 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.215169 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.215187 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:57Z","lastTransitionTime":"2025-09-30T13:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.317783 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.317849 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.317874 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.317905 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.317930 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:57Z","lastTransitionTime":"2025-09-30T13:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.420118 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.420181 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.420199 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.420215 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.420225 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:57Z","lastTransitionTime":"2025-09-30T13:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.523256 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.523299 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.523308 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.523323 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.523333 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:57Z","lastTransitionTime":"2025-09-30T13:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.524528 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.524583 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.524596 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.524611 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.524623 4840 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T13:57:57Z","lastTransitionTime":"2025-09-30T13:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.569852 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-nrv6j"] Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.570273 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nrv6j" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.572073 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.572266 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.572348 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.572405 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.651451 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/97581540-9409-41d7-b058-9c79f5be7c78-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-nrv6j\" (UID: \"97581540-9409-41d7-b058-9c79f5be7c78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nrv6j" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.651503 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97581540-9409-41d7-b058-9c79f5be7c78-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-nrv6j\" (UID: \"97581540-9409-41d7-b058-9c79f5be7c78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nrv6j" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.651739 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/97581540-9409-41d7-b058-9c79f5be7c78-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-nrv6j\" (UID: \"97581540-9409-41d7-b058-9c79f5be7c78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nrv6j" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.651790 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/97581540-9409-41d7-b058-9c79f5be7c78-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-nrv6j\" (UID: \"97581540-9409-41d7-b058-9c79f5be7c78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nrv6j" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.651971 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/97581540-9409-41d7-b058-9c79f5be7c78-service-ca\") pod \"cluster-version-operator-5c965bbfc6-nrv6j\" (UID: \"97581540-9409-41d7-b058-9c79f5be7c78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nrv6j" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.753079 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/97581540-9409-41d7-b058-9c79f5be7c78-service-ca\") pod \"cluster-version-operator-5c965bbfc6-nrv6j\" (UID: \"97581540-9409-41d7-b058-9c79f5be7c78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nrv6j" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.753131 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/97581540-9409-41d7-b058-9c79f5be7c78-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-nrv6j\" (UID: \"97581540-9409-41d7-b058-9c79f5be7c78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nrv6j" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.753153 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97581540-9409-41d7-b058-9c79f5be7c78-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-nrv6j\" (UID: \"97581540-9409-41d7-b058-9c79f5be7c78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nrv6j" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.753173 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/97581540-9409-41d7-b058-9c79f5be7c78-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-nrv6j\" (UID: \"97581540-9409-41d7-b058-9c79f5be7c78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nrv6j" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.753190 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/97581540-9409-41d7-b058-9c79f5be7c78-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-nrv6j\" (UID: \"97581540-9409-41d7-b058-9c79f5be7c78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nrv6j" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.753292 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/97581540-9409-41d7-b058-9c79f5be7c78-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-nrv6j\" (UID: \"97581540-9409-41d7-b058-9c79f5be7c78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nrv6j" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.753386 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/97581540-9409-41d7-b058-9c79f5be7c78-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-nrv6j\" (UID: \"97581540-9409-41d7-b058-9c79f5be7c78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nrv6j" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.754007 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/97581540-9409-41d7-b058-9c79f5be7c78-service-ca\") pod \"cluster-version-operator-5c965bbfc6-nrv6j\" (UID: \"97581540-9409-41d7-b058-9c79f5be7c78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nrv6j" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.759724 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97581540-9409-41d7-b058-9c79f5be7c78-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-nrv6j\" (UID: \"97581540-9409-41d7-b058-9c79f5be7c78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nrv6j" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.769046 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/97581540-9409-41d7-b058-9c79f5be7c78-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-nrv6j\" (UID: \"97581540-9409-41d7-b058-9c79f5be7c78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nrv6j" Sep 30 13:57:57 crc kubenswrapper[4840]: I0930 13:57:57.887182 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nrv6j" Sep 30 13:57:58 crc kubenswrapper[4840]: I0930 13:57:58.116186 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:57:58 crc kubenswrapper[4840]: E0930 13:57:58.116628 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:57:58 crc kubenswrapper[4840]: I0930 13:57:58.664352 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nrv6j" event={"ID":"97581540-9409-41d7-b058-9c79f5be7c78","Type":"ContainerStarted","Data":"aa9045d7a911d7e545d8c14d171ed3cdb4e7e3b65295728daa7c41921f8f70cb"} Sep 30 13:57:58 crc kubenswrapper[4840]: I0930 13:57:58.664402 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nrv6j" event={"ID":"97581540-9409-41d7-b058-9c79f5be7c78","Type":"ContainerStarted","Data":"a36bcb551e76d99b6e2c5f8052a034b87bd7a61b445a634c366f414051720de9"} Sep 30 13:57:58 crc kubenswrapper[4840]: I0930 13:57:58.680098 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nrv6j" podStartSLOduration=84.680070936 podStartE2EDuration="1m24.680070936s" podCreationTimestamp="2025-09-30 13:56:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:57:58.678796582 +0000 UTC m=+107.307883005" watchObservedRunningTime="2025-09-30 13:57:58.680070936 +0000 UTC m=+107.309157379" Sep 30 13:57:59 crc kubenswrapper[4840]: I0930 13:57:59.116261 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:57:59 crc kubenswrapper[4840]: I0930 13:57:59.116283 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:57:59 crc kubenswrapper[4840]: I0930 13:57:59.116369 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:57:59 crc kubenswrapper[4840]: E0930 13:57:59.116588 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:57:59 crc kubenswrapper[4840]: E0930 13:57:59.116669 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:57:59 crc kubenswrapper[4840]: E0930 13:57:59.116758 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:58:00 crc kubenswrapper[4840]: I0930 13:58:00.116053 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:58:00 crc kubenswrapper[4840]: E0930 13:58:00.116261 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:58:01 crc kubenswrapper[4840]: I0930 13:58:01.116504 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:58:01 crc kubenswrapper[4840]: I0930 13:58:01.116604 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:58:01 crc kubenswrapper[4840]: I0930 13:58:01.116690 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:58:01 crc kubenswrapper[4840]: E0930 13:58:01.116684 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:58:01 crc kubenswrapper[4840]: E0930 13:58:01.117099 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:58:01 crc kubenswrapper[4840]: E0930 13:58:01.117213 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:58:02 crc kubenswrapper[4840]: I0930 13:58:02.117000 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:58:02 crc kubenswrapper[4840]: E0930 13:58:02.117166 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:58:03 crc kubenswrapper[4840]: I0930 13:58:03.116088 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:58:03 crc kubenswrapper[4840]: I0930 13:58:03.116122 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:58:03 crc kubenswrapper[4840]: E0930 13:58:03.116235 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:58:03 crc kubenswrapper[4840]: I0930 13:58:03.116085 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:58:03 crc kubenswrapper[4840]: E0930 13:58:03.116340 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:58:03 crc kubenswrapper[4840]: E0930 13:58:03.116578 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:58:04 crc kubenswrapper[4840]: I0930 13:58:04.116222 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:58:04 crc kubenswrapper[4840]: E0930 13:58:04.116447 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:58:05 crc kubenswrapper[4840]: I0930 13:58:05.116053 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:58:05 crc kubenswrapper[4840]: I0930 13:58:05.116130 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:58:05 crc kubenswrapper[4840]: I0930 13:58:05.116139 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:58:05 crc kubenswrapper[4840]: E0930 13:58:05.116295 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:58:05 crc kubenswrapper[4840]: E0930 13:58:05.116360 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:58:05 crc kubenswrapper[4840]: E0930 13:58:05.116457 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:58:05 crc kubenswrapper[4840]: I0930 13:58:05.117507 4840 scope.go:117] "RemoveContainer" containerID="109628b3a230ce972e436d712479c6f8071e34a983f83d79555ffea902ca2e55" Sep 30 13:58:05 crc kubenswrapper[4840]: E0930 13:58:05.117781 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-kfx69_openshift-ovn-kubernetes(2ff48c28-d076-46e8-a93f-9630989f81e8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" Sep 30 13:58:06 crc kubenswrapper[4840]: I0930 13:58:06.115588 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:58:06 crc kubenswrapper[4840]: E0930 13:58:06.115768 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:58:07 crc kubenswrapper[4840]: I0930 13:58:07.115689 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:58:07 crc kubenswrapper[4840]: E0930 13:58:07.115814 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:58:07 crc kubenswrapper[4840]: I0930 13:58:07.115836 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:58:07 crc kubenswrapper[4840]: E0930 13:58:07.115896 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:58:07 crc kubenswrapper[4840]: I0930 13:58:07.115826 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:58:07 crc kubenswrapper[4840]: E0930 13:58:07.115956 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:58:08 crc kubenswrapper[4840]: I0930 13:58:08.115841 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:58:08 crc kubenswrapper[4840]: E0930 13:58:08.115988 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:58:09 crc kubenswrapper[4840]: I0930 13:58:09.115792 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:58:09 crc kubenswrapper[4840]: I0930 13:58:09.115808 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:58:09 crc kubenswrapper[4840]: I0930 13:58:09.115848 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:58:09 crc kubenswrapper[4840]: E0930 13:58:09.116336 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:58:09 crc kubenswrapper[4840]: E0930 13:58:09.116489 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:58:09 crc kubenswrapper[4840]: E0930 13:58:09.116201 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:58:10 crc kubenswrapper[4840]: I0930 13:58:10.115968 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:58:10 crc kubenswrapper[4840]: E0930 13:58:10.116137 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:58:11 crc kubenswrapper[4840]: I0930 13:58:11.116348 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:58:11 crc kubenswrapper[4840]: E0930 13:58:11.116794 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:58:11 crc kubenswrapper[4840]: I0930 13:58:11.116667 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:58:11 crc kubenswrapper[4840]: I0930 13:58:11.116631 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:58:11 crc kubenswrapper[4840]: E0930 13:58:11.116877 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:58:11 crc kubenswrapper[4840]: E0930 13:58:11.117032 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:58:12 crc kubenswrapper[4840]: E0930 13:58:12.031080 4840 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Sep 30 13:58:12 crc kubenswrapper[4840]: I0930 13:58:12.115961 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:58:12 crc kubenswrapper[4840]: E0930 13:58:12.117122 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:58:12 crc kubenswrapper[4840]: E0930 13:58:12.255184 4840 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Sep 30 13:58:13 crc kubenswrapper[4840]: I0930 13:58:13.115752 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:58:13 crc kubenswrapper[4840]: I0930 13:58:13.115758 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:58:13 crc kubenswrapper[4840]: E0930 13:58:13.115955 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:58:13 crc kubenswrapper[4840]: E0930 13:58:13.115875 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:58:13 crc kubenswrapper[4840]: I0930 13:58:13.116127 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:58:13 crc kubenswrapper[4840]: E0930 13:58:13.116334 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:58:14 crc kubenswrapper[4840]: I0930 13:58:14.115526 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:58:14 crc kubenswrapper[4840]: E0930 13:58:14.115690 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:58:15 crc kubenswrapper[4840]: I0930 13:58:15.115607 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:58:15 crc kubenswrapper[4840]: I0930 13:58:15.115638 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:58:15 crc kubenswrapper[4840]: I0930 13:58:15.115713 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:58:15 crc kubenswrapper[4840]: E0930 13:58:15.115749 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:58:15 crc kubenswrapper[4840]: E0930 13:58:15.115882 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:58:15 crc kubenswrapper[4840]: E0930 13:58:15.115996 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:58:15 crc kubenswrapper[4840]: I0930 13:58:15.720825 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bwvl2_cbe233ee-1ea7-433e-a53a-e4a668f739ee/kube-multus/1.log" Sep 30 13:58:15 crc kubenswrapper[4840]: I0930 13:58:15.721453 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bwvl2_cbe233ee-1ea7-433e-a53a-e4a668f739ee/kube-multus/0.log" Sep 30 13:58:15 crc kubenswrapper[4840]: I0930 13:58:15.721609 4840 generic.go:334] "Generic (PLEG): container finished" podID="cbe233ee-1ea7-433e-a53a-e4a668f739ee" containerID="a3eb5795ae35b00eba9462e95fdba89c2dcf0442843c32ec05edd710dfe507fe" exitCode=1 Sep 30 13:58:15 crc kubenswrapper[4840]: I0930 13:58:15.721676 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bwvl2" event={"ID":"cbe233ee-1ea7-433e-a53a-e4a668f739ee","Type":"ContainerDied","Data":"a3eb5795ae35b00eba9462e95fdba89c2dcf0442843c32ec05edd710dfe507fe"} Sep 30 13:58:15 crc kubenswrapper[4840]: I0930 13:58:15.721804 4840 scope.go:117] "RemoveContainer" containerID="1d94417d8a6678a5043042aae431a3a8f0d4456a00ea3e80e3fc92bf5a819c2a" Sep 30 13:58:15 crc kubenswrapper[4840]: I0930 13:58:15.722733 4840 scope.go:117] "RemoveContainer" containerID="a3eb5795ae35b00eba9462e95fdba89c2dcf0442843c32ec05edd710dfe507fe" Sep 30 13:58:15 crc kubenswrapper[4840]: E0930 13:58:15.723154 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-bwvl2_openshift-multus(cbe233ee-1ea7-433e-a53a-e4a668f739ee)\"" pod="openshift-multus/multus-bwvl2" podUID="cbe233ee-1ea7-433e-a53a-e4a668f739ee" Sep 30 13:58:16 crc kubenswrapper[4840]: I0930 13:58:16.116478 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:58:16 crc kubenswrapper[4840]: E0930 13:58:16.116918 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:58:16 crc kubenswrapper[4840]: I0930 13:58:16.117637 4840 scope.go:117] "RemoveContainer" containerID="109628b3a230ce972e436d712479c6f8071e34a983f83d79555ffea902ca2e55" Sep 30 13:58:16 crc kubenswrapper[4840]: E0930 13:58:16.117843 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-kfx69_openshift-ovn-kubernetes(2ff48c28-d076-46e8-a93f-9630989f81e8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" Sep 30 13:58:16 crc kubenswrapper[4840]: I0930 13:58:16.727877 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bwvl2_cbe233ee-1ea7-433e-a53a-e4a668f739ee/kube-multus/1.log" Sep 30 13:58:17 crc kubenswrapper[4840]: I0930 13:58:17.116201 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:58:17 crc kubenswrapper[4840]: I0930 13:58:17.116250 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:58:17 crc kubenswrapper[4840]: I0930 13:58:17.116321 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:58:17 crc kubenswrapper[4840]: E0930 13:58:17.116393 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:58:17 crc kubenswrapper[4840]: E0930 13:58:17.116494 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:58:17 crc kubenswrapper[4840]: E0930 13:58:17.116537 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:58:17 crc kubenswrapper[4840]: E0930 13:58:17.256147 4840 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Sep 30 13:58:18 crc kubenswrapper[4840]: I0930 13:58:18.115535 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:58:18 crc kubenswrapper[4840]: E0930 13:58:18.115703 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:58:19 crc kubenswrapper[4840]: I0930 13:58:19.116622 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:58:19 crc kubenswrapper[4840]: I0930 13:58:19.116686 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:58:19 crc kubenswrapper[4840]: I0930 13:58:19.116620 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:58:19 crc kubenswrapper[4840]: E0930 13:58:19.116772 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:58:19 crc kubenswrapper[4840]: E0930 13:58:19.117160 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:58:19 crc kubenswrapper[4840]: E0930 13:58:19.117926 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:58:20 crc kubenswrapper[4840]: I0930 13:58:20.115723 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:58:20 crc kubenswrapper[4840]: E0930 13:58:20.115863 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:58:21 crc kubenswrapper[4840]: I0930 13:58:21.115895 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:58:21 crc kubenswrapper[4840]: I0930 13:58:21.115934 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:58:21 crc kubenswrapper[4840]: E0930 13:58:21.116048 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:58:21 crc kubenswrapper[4840]: E0930 13:58:21.116184 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:58:21 crc kubenswrapper[4840]: I0930 13:58:21.116519 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:58:21 crc kubenswrapper[4840]: E0930 13:58:21.116654 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:58:22 crc kubenswrapper[4840]: I0930 13:58:22.116149 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:58:22 crc kubenswrapper[4840]: E0930 13:58:22.117597 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:58:22 crc kubenswrapper[4840]: E0930 13:58:22.256935 4840 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Sep 30 13:58:23 crc kubenswrapper[4840]: I0930 13:58:23.115447 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:58:23 crc kubenswrapper[4840]: I0930 13:58:23.115499 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:58:23 crc kubenswrapper[4840]: E0930 13:58:23.115726 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:58:23 crc kubenswrapper[4840]: E0930 13:58:23.115835 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:58:23 crc kubenswrapper[4840]: I0930 13:58:23.115959 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:58:23 crc kubenswrapper[4840]: E0930 13:58:23.116068 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:58:24 crc kubenswrapper[4840]: I0930 13:58:24.115692 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:58:24 crc kubenswrapper[4840]: E0930 13:58:24.115844 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:58:25 crc kubenswrapper[4840]: I0930 13:58:25.115514 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:58:25 crc kubenswrapper[4840]: I0930 13:58:25.115543 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:58:25 crc kubenswrapper[4840]: I0930 13:58:25.115678 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:58:25 crc kubenswrapper[4840]: E0930 13:58:25.115761 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:58:25 crc kubenswrapper[4840]: E0930 13:58:25.115863 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:58:25 crc kubenswrapper[4840]: E0930 13:58:25.115952 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:58:26 crc kubenswrapper[4840]: I0930 13:58:26.116174 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:58:26 crc kubenswrapper[4840]: E0930 13:58:26.116308 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:58:27 crc kubenswrapper[4840]: I0930 13:58:27.116004 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:58:27 crc kubenswrapper[4840]: I0930 13:58:27.116105 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:58:27 crc kubenswrapper[4840]: E0930 13:58:27.116212 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:58:27 crc kubenswrapper[4840]: I0930 13:58:27.116302 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:58:27 crc kubenswrapper[4840]: E0930 13:58:27.116501 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:58:27 crc kubenswrapper[4840]: E0930 13:58:27.116621 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:58:27 crc kubenswrapper[4840]: E0930 13:58:27.258858 4840 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Sep 30 13:58:28 crc kubenswrapper[4840]: I0930 13:58:28.115585 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:58:28 crc kubenswrapper[4840]: E0930 13:58:28.115739 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:58:29 crc kubenswrapper[4840]: I0930 13:58:29.115585 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:58:29 crc kubenswrapper[4840]: I0930 13:58:29.115656 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:58:29 crc kubenswrapper[4840]: I0930 13:58:29.115677 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:58:29 crc kubenswrapper[4840]: E0930 13:58:29.115814 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:58:29 crc kubenswrapper[4840]: I0930 13:58:29.115977 4840 scope.go:117] "RemoveContainer" containerID="a3eb5795ae35b00eba9462e95fdba89c2dcf0442843c32ec05edd710dfe507fe" Sep 30 13:58:29 crc kubenswrapper[4840]: E0930 13:58:29.116032 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:58:29 crc kubenswrapper[4840]: E0930 13:58:29.116081 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:58:29 crc kubenswrapper[4840]: I0930 13:58:29.775483 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bwvl2_cbe233ee-1ea7-433e-a53a-e4a668f739ee/kube-multus/1.log" Sep 30 13:58:29 crc kubenswrapper[4840]: I0930 13:58:29.775879 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bwvl2" event={"ID":"cbe233ee-1ea7-433e-a53a-e4a668f739ee","Type":"ContainerStarted","Data":"16d3d0c48c4c35832f75f43479de2114774ae32ec74b4f01cc9e8a63dee564c6"} Sep 30 13:58:30 crc kubenswrapper[4840]: I0930 13:58:30.115534 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:58:30 crc kubenswrapper[4840]: E0930 13:58:30.115770 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:58:31 crc kubenswrapper[4840]: I0930 13:58:31.116276 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:58:31 crc kubenswrapper[4840]: I0930 13:58:31.116348 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:58:31 crc kubenswrapper[4840]: I0930 13:58:31.116427 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:58:31 crc kubenswrapper[4840]: E0930 13:58:31.116438 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:58:31 crc kubenswrapper[4840]: E0930 13:58:31.116610 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:58:31 crc kubenswrapper[4840]: E0930 13:58:31.116808 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:58:31 crc kubenswrapper[4840]: I0930 13:58:31.117594 4840 scope.go:117] "RemoveContainer" containerID="109628b3a230ce972e436d712479c6f8071e34a983f83d79555ffea902ca2e55" Sep 30 13:58:31 crc kubenswrapper[4840]: I0930 13:58:31.785132 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kfx69_2ff48c28-d076-46e8-a93f-9630989f81e8/ovnkube-controller/3.log" Sep 30 13:58:31 crc kubenswrapper[4840]: I0930 13:58:31.788570 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" event={"ID":"2ff48c28-d076-46e8-a93f-9630989f81e8","Type":"ContainerStarted","Data":"6fdd66531df81c022d7053dc10bb0f7c09da2dabfacf7993d7439a395f2f453b"} Sep 30 13:58:31 crc kubenswrapper[4840]: I0930 13:58:31.789090 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:58:31 crc kubenswrapper[4840]: I0930 13:58:31.818196 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" podStartSLOduration=116.818169041 podStartE2EDuration="1m56.818169041s" podCreationTimestamp="2025-09-30 13:56:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:31.817451192 +0000 UTC m=+140.446537635" watchObservedRunningTime="2025-09-30 13:58:31.818169041 +0000 UTC m=+140.447255464" Sep 30 13:58:31 crc kubenswrapper[4840]: I0930 13:58:31.887913 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-4gp5x"] Sep 30 13:58:31 crc kubenswrapper[4840]: I0930 13:58:31.888045 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:58:31 crc kubenswrapper[4840]: E0930 13:58:31.888153 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:58:32 crc kubenswrapper[4840]: I0930 13:58:32.115652 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:58:32 crc kubenswrapper[4840]: E0930 13:58:32.120035 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:58:32 crc kubenswrapper[4840]: E0930 13:58:32.259785 4840 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Sep 30 13:58:33 crc kubenswrapper[4840]: I0930 13:58:33.115852 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:58:33 crc kubenswrapper[4840]: I0930 13:58:33.115868 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:58:33 crc kubenswrapper[4840]: E0930 13:58:33.115986 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:58:33 crc kubenswrapper[4840]: I0930 13:58:33.116342 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:58:33 crc kubenswrapper[4840]: E0930 13:58:33.116355 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:58:33 crc kubenswrapper[4840]: E0930 13:58:33.116497 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:58:34 crc kubenswrapper[4840]: I0930 13:58:34.115733 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:58:34 crc kubenswrapper[4840]: E0930 13:58:34.115852 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:58:35 crc kubenswrapper[4840]: I0930 13:58:35.116302 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:58:35 crc kubenswrapper[4840]: E0930 13:58:35.116464 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:58:35 crc kubenswrapper[4840]: I0930 13:58:35.116934 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:58:35 crc kubenswrapper[4840]: I0930 13:58:35.116976 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:58:35 crc kubenswrapper[4840]: E0930 13:58:35.117026 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:58:35 crc kubenswrapper[4840]: E0930 13:58:35.117132 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:58:36 crc kubenswrapper[4840]: I0930 13:58:36.115650 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:58:36 crc kubenswrapper[4840]: E0930 13:58:36.115848 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 13:58:37 crc kubenswrapper[4840]: I0930 13:58:37.116484 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:58:37 crc kubenswrapper[4840]: I0930 13:58:37.116484 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:58:37 crc kubenswrapper[4840]: I0930 13:58:37.116516 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:58:37 crc kubenswrapper[4840]: E0930 13:58:37.116703 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 13:58:37 crc kubenswrapper[4840]: E0930 13:58:37.116958 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4gp5x" podUID="1491f559-bc12-4afd-a40c-4eaa40d920a8" Sep 30 13:58:37 crc kubenswrapper[4840]: E0930 13:58:37.117097 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.116507 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.118411 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.118925 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.181003 4840 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.212395 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-wdjfz"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.212847 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-wdjfz" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.213125 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-qjxnf"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.213892 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-qjxnf" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.214272 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftgw2"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.214827 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftgw2" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.215106 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.215478 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9rnrg"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.216034 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9rnrg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.220153 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.220186 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.223061 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.224262 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.224606 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.225063 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-v676n"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.225139 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.225503 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.225534 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-v676n" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.225703 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.225866 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.226628 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.226689 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.226806 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.227035 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.229118 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.231540 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-kdwqh"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.232053 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-kdwqh" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.232079 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.232139 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.232138 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.232146 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.232219 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.232236 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.232255 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.232494 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.233937 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.234138 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.234813 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.235006 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.235015 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.235129 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.237291 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-p76w8"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.237942 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p76w8" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.238355 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-zbnls"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.239159 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-zbnls" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.240629 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-9s79z"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.241190 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9s79z" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.244635 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ncblm"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.245142 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.248834 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.248978 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.266168 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m2bq7"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.267910 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.268462 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.268730 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.269030 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m2bq7" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.270234 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.271199 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.285304 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.285476 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-lk4jg"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.286112 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-jfnms"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.286324 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-dckvx"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.286609 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-dckvx" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.286728 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.286809 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.286865 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-jfnms" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.286917 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.287036 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.287104 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.287189 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.287272 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.287457 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.287494 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.287626 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.287748 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.287824 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.287890 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.288046 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.288118 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.288920 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.289141 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.289593 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.290017 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.290163 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.290407 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.293679 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.295072 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.295217 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.295680 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.296627 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sc7xq"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.300492 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9dd57b46-77fc-4b8a-9fc1-a50084165914-trusted-ca-bundle\") pod \"console-f9d7485db-qjxnf\" (UID: \"9dd57b46-77fc-4b8a-9fc1-a50084165914\") " pod="openshift-console/console-f9d7485db-qjxnf" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.300539 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2z4f\" (UniqueName: \"kubernetes.io/projected/bb0b27d1-5cb0-4a2c-95d7-012bc0a2d071-kube-api-access-v2z4f\") pod \"authentication-operator-69f744f599-v676n\" (UID: \"bb0b27d1-5cb0-4a2c-95d7-012bc0a2d071\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v676n" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.300586 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/661771aa-bd40-40fd-8a59-dc3b8156e5c5-client-ca\") pod \"controller-manager-879f6c89f-wdjfz\" (UID: \"661771aa-bd40-40fd-8a59-dc3b8156e5c5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wdjfz" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.300608 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvvvw\" (UniqueName: \"kubernetes.io/projected/2842cd75-1ff5-4a3f-b7a0-8205388f4a17-kube-api-access-vvvvw\") pod \"route-controller-manager-6576b87f9c-ftgw2\" (UID: \"2842cd75-1ff5-4a3f-b7a0-8205388f4a17\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftgw2" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.300630 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9dd57b46-77fc-4b8a-9fc1-a50084165914-console-serving-cert\") pod \"console-f9d7485db-qjxnf\" (UID: \"9dd57b46-77fc-4b8a-9fc1-a50084165914\") " pod="openshift-console/console-f9d7485db-qjxnf" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.300674 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5b44f815-36aa-42d8-9c3e-fb8d956c05c2-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-9rnrg\" (UID: \"5b44f815-36aa-42d8-9c3e-fb8d956c05c2\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9rnrg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.300697 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chhwt\" (UniqueName: \"kubernetes.io/projected/661771aa-bd40-40fd-8a59-dc3b8156e5c5-kube-api-access-chhwt\") pod \"controller-manager-879f6c89f-wdjfz\" (UID: \"661771aa-bd40-40fd-8a59-dc3b8156e5c5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wdjfz" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.300722 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2842cd75-1ff5-4a3f-b7a0-8205388f4a17-serving-cert\") pod \"route-controller-manager-6576b87f9c-ftgw2\" (UID: \"2842cd75-1ff5-4a3f-b7a0-8205388f4a17\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftgw2" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.300744 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb0b27d1-5cb0-4a2c-95d7-012bc0a2d071-service-ca-bundle\") pod \"authentication-operator-69f744f599-v676n\" (UID: \"bb0b27d1-5cb0-4a2c-95d7-012bc0a2d071\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v676n" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.300768 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2842cd75-1ff5-4a3f-b7a0-8205388f4a17-config\") pod \"route-controller-manager-6576b87f9c-ftgw2\" (UID: \"2842cd75-1ff5-4a3f-b7a0-8205388f4a17\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftgw2" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.300787 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/661771aa-bd40-40fd-8a59-dc3b8156e5c5-serving-cert\") pod \"controller-manager-879f6c89f-wdjfz\" (UID: \"661771aa-bd40-40fd-8a59-dc3b8156e5c5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wdjfz" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.300812 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb0b27d1-5cb0-4a2c-95d7-012bc0a2d071-config\") pod \"authentication-operator-69f744f599-v676n\" (UID: \"bb0b27d1-5cb0-4a2c-95d7-012bc0a2d071\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v676n" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.300835 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/661771aa-bd40-40fd-8a59-dc3b8156e5c5-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-wdjfz\" (UID: \"661771aa-bd40-40fd-8a59-dc3b8156e5c5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wdjfz" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.300879 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9dd57b46-77fc-4b8a-9fc1-a50084165914-console-config\") pod \"console-f9d7485db-qjxnf\" (UID: \"9dd57b46-77fc-4b8a-9fc1-a50084165914\") " pod="openshift-console/console-f9d7485db-qjxnf" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.300902 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9dd57b46-77fc-4b8a-9fc1-a50084165914-service-ca\") pod \"console-f9d7485db-qjxnf\" (UID: \"9dd57b46-77fc-4b8a-9fc1-a50084165914\") " pod="openshift-console/console-f9d7485db-qjxnf" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.300927 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9dd57b46-77fc-4b8a-9fc1-a50084165914-oauth-serving-cert\") pod \"console-f9d7485db-qjxnf\" (UID: \"9dd57b46-77fc-4b8a-9fc1-a50084165914\") " pod="openshift-console/console-f9d7485db-qjxnf" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.300960 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb0b27d1-5cb0-4a2c-95d7-012bc0a2d071-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-v676n\" (UID: \"bb0b27d1-5cb0-4a2c-95d7-012bc0a2d071\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v676n" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.300981 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2842cd75-1ff5-4a3f-b7a0-8205388f4a17-client-ca\") pod \"route-controller-manager-6576b87f9c-ftgw2\" (UID: \"2842cd75-1ff5-4a3f-b7a0-8205388f4a17\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftgw2" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.301002 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9dd57b46-77fc-4b8a-9fc1-a50084165914-console-oauth-config\") pod \"console-f9d7485db-qjxnf\" (UID: \"9dd57b46-77fc-4b8a-9fc1-a50084165914\") " pod="openshift-console/console-f9d7485db-qjxnf" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.301021 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpmvq\" (UniqueName: \"kubernetes.io/projected/9dd57b46-77fc-4b8a-9fc1-a50084165914-kube-api-access-cpmvq\") pod \"console-f9d7485db-qjxnf\" (UID: \"9dd57b46-77fc-4b8a-9fc1-a50084165914\") " pod="openshift-console/console-f9d7485db-qjxnf" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.301044 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/661771aa-bd40-40fd-8a59-dc3b8156e5c5-config\") pod \"controller-manager-879f6c89f-wdjfz\" (UID: \"661771aa-bd40-40fd-8a59-dc3b8156e5c5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wdjfz" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.301086 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qr4f5\" (UniqueName: \"kubernetes.io/projected/5b44f815-36aa-42d8-9c3e-fb8d956c05c2-kube-api-access-qr4f5\") pod \"cluster-samples-operator-665b6dd947-9rnrg\" (UID: \"5b44f815-36aa-42d8-9c3e-fb8d956c05c2\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9rnrg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.301109 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb0b27d1-5cb0-4a2c-95d7-012bc0a2d071-serving-cert\") pod \"authentication-operator-69f744f599-v676n\" (UID: \"bb0b27d1-5cb0-4a2c-95d7-012bc0a2d071\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v676n" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.301131 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmp25\" (UniqueName: \"kubernetes.io/projected/dde0d17e-6bd0-4acb-b8b8-a8b0bc988a6e-kube-api-access-mmp25\") pod \"downloads-7954f5f757-kdwqh\" (UID: \"dde0d17e-6bd0-4acb-b8b8-a8b0bc988a6e\") " pod="openshift-console/downloads-7954f5f757-kdwqh" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.301249 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sc7xq" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.302966 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.303208 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.303294 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.303385 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.303385 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.303571 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.303614 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpb68"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.303633 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.303673 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.303704 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.303721 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.303768 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.303655 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.303876 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.303893 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.303919 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.304047 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.304106 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpb68" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.304275 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.304380 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.304403 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.304440 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.304387 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.304483 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.304513 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.304547 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.306026 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-l47x7"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.306614 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-bf4gt"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.307021 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bf4gt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.307085 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-l47x7" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.308954 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-s9lx8"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.309692 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-s9lx8" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.314356 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.314905 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.315381 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.318391 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vckwz"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.318917 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.319077 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.319316 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vckwz" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.319765 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.320145 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.320157 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.320188 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4942b"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.323095 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.323514 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.324445 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4942b" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.333735 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-bgjpl"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.344438 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.358455 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.362235 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.366270 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.371166 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.371219 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9szj8"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.371584 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-lzwls"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.371850 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kkmmf"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.372130 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9szj8" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.372155 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bgjpl" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.372287 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.372542 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.372703 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-ptz4x"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.373162 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-ptz4x" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.373342 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kkmmf" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.375653 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-wdjfz"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.376052 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.376083 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.378826 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.379007 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftgw2"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.379065 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-jslz5"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.381189 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n2cvx"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.381476 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-jslz5" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.382124 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.382688 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n2cvx" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.385980 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-czbs2"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.386835 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-czbs2" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.390594 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-dksqg"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.391307 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dksqg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.392943 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-snznw"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.393771 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-snznw" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.395038 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.395820 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z866q"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.396450 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z866q" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.396596 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s72m9"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.397432 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s72m9" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.400109 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-w4vtf"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.400792 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-8h224"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.401440 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-68gcw"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.401520 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-w4vtf" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.401620 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2b720224-3e65-4443-a056-4e8fd226bcea-audit-policies\") pod \"apiserver-7bbb656c7d-bf4gt\" (UID: \"2b720224-3e65-4443-a056-4e8fd226bcea\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bf4gt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.401654 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e8eb0c45-6539-478d-96ff-1d43d5d9978b-auth-proxy-config\") pod \"machine-approver-56656f9798-p76w8\" (UID: \"e8eb0c45-6539-478d-96ff-1d43d5d9978b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p76w8" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.401684 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/661771aa-bd40-40fd-8a59-dc3b8156e5c5-config\") pod \"controller-manager-879f6c89f-wdjfz\" (UID: \"661771aa-bd40-40fd-8a59-dc3b8156e5c5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wdjfz" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.401704 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0ba2ddf1-a9b7-448a-9e48-cdc9e80f3873-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-4942b\" (UID: \"0ba2ddf1-a9b7-448a-9e48-cdc9e80f3873\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4942b" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.401722 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmm2z\" (UniqueName: \"kubernetes.io/projected/8ec02940-3b96-41aa-8624-8338ee544fe1-kube-api-access-fmm2z\") pod \"apiserver-76f77b778f-lk4jg\" (UID: \"8ec02940-3b96-41aa-8624-8338ee544fe1\") " pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.401753 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0ba2ddf1-a9b7-448a-9e48-cdc9e80f3873-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-4942b\" (UID: \"0ba2ddf1-a9b7-448a-9e48-cdc9e80f3873\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4942b" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.401771 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7brf9\" (UniqueName: \"kubernetes.io/projected/e8eb0c45-6539-478d-96ff-1d43d5d9978b-kube-api-access-7brf9\") pod \"machine-approver-56656f9798-p76w8\" (UID: \"e8eb0c45-6539-478d-96ff-1d43d5d9978b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p76w8" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.401787 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8dpt\" (UniqueName: \"kubernetes.io/projected/ea0a1a12-942c-4ca3-b649-5ad87af478c8-kube-api-access-d8dpt\") pod \"openshift-controller-manager-operator-756b6f6bc6-sc7xq\" (UID: \"ea0a1a12-942c-4ca3-b649-5ad87af478c8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sc7xq" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.401807 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvhjm\" (UniqueName: \"kubernetes.io/projected/4643a53d-3236-4642-9206-603d20a15129-kube-api-access-bvhjm\") pod \"etcd-operator-b45778765-dckvx\" (UID: \"4643a53d-3236-4642-9206-603d20a15129\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dckvx" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.401833 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4643a53d-3236-4642-9206-603d20a15129-serving-cert\") pod \"etcd-operator-b45778765-dckvx\" (UID: \"4643a53d-3236-4642-9206-603d20a15129\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dckvx" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.401852 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmp25\" (UniqueName: \"kubernetes.io/projected/dde0d17e-6bd0-4acb-b8b8-a8b0bc988a6e-kube-api-access-mmp25\") pod \"downloads-7954f5f757-kdwqh\" (UID: \"dde0d17e-6bd0-4acb-b8b8-a8b0bc988a6e\") " pod="openshift-console/downloads-7954f5f757-kdwqh" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.401874 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8d0002f3-3315-4c27-bc39-2e5989b67d6b-serving-cert\") pod \"openshift-config-operator-7777fb866f-9s79z\" (UID: \"8d0002f3-3315-4c27-bc39-2e5989b67d6b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9s79z" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.401893 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qr4f5\" (UniqueName: \"kubernetes.io/projected/5b44f815-36aa-42d8-9c3e-fb8d956c05c2-kube-api-access-qr4f5\") pod \"cluster-samples-operator-665b6dd947-9rnrg\" (UID: \"5b44f815-36aa-42d8-9c3e-fb8d956c05c2\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9rnrg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.401913 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb0b27d1-5cb0-4a2c-95d7-012bc0a2d071-serving-cert\") pod \"authentication-operator-69f744f599-v676n\" (UID: \"bb0b27d1-5cb0-4a2c-95d7-012bc0a2d071\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v676n" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.401927 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2b720224-3e65-4443-a056-4e8fd226bcea-etcd-client\") pod \"apiserver-7bbb656c7d-bf4gt\" (UID: \"2b720224-3e65-4443-a056-4e8fd226bcea\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bf4gt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.401943 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8ec02940-3b96-41aa-8624-8338ee544fe1-etcd-client\") pod \"apiserver-76f77b778f-lk4jg\" (UID: \"8ec02940-3b96-41aa-8624-8338ee544fe1\") " pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.401957 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46efdeea-a4d2-4ff5-898d-9ee516105cf4-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vckwz\" (UID: \"46efdeea-a4d2-4ff5-898d-9ee516105cf4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vckwz" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.401976 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.401993 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8d0002f3-3315-4c27-bc39-2e5989b67d6b-available-featuregates\") pod \"openshift-config-operator-7777fb866f-9s79z\" (UID: \"8d0002f3-3315-4c27-bc39-2e5989b67d6b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9s79z" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402011 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9dd57b46-77fc-4b8a-9fc1-a50084165914-trusted-ca-bundle\") pod \"console-f9d7485db-qjxnf\" (UID: \"9dd57b46-77fc-4b8a-9fc1-a50084165914\") " pod="openshift-console/console-f9d7485db-qjxnf" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402027 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/8ec02940-3b96-41aa-8624-8338ee544fe1-audit\") pod \"apiserver-76f77b778f-lk4jg\" (UID: \"8ec02940-3b96-41aa-8624-8338ee544fe1\") " pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402042 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2b720224-3e65-4443-a056-4e8fd226bcea-audit-dir\") pod \"apiserver-7bbb656c7d-bf4gt\" (UID: \"2b720224-3e65-4443-a056-4e8fd226bcea\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bf4gt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402051 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-68gcw" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402060 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmrwg\" (UniqueName: \"kubernetes.io/projected/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-kube-api-access-qmrwg\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402076 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdp5r\" (UniqueName: \"kubernetes.io/projected/8d0002f3-3315-4c27-bc39-2e5989b67d6b-kube-api-access-jdp5r\") pod \"openshift-config-operator-7777fb866f-9s79z\" (UID: \"8d0002f3-3315-4c27-bc39-2e5989b67d6b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9s79z" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402092 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2z4f\" (UniqueName: \"kubernetes.io/projected/bb0b27d1-5cb0-4a2c-95d7-012bc0a2d071-kube-api-access-v2z4f\") pod \"authentication-operator-69f744f599-v676n\" (UID: \"bb0b27d1-5cb0-4a2c-95d7-012bc0a2d071\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v676n" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402110 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gb9h8\" (UniqueName: \"kubernetes.io/projected/d0cf2421-0839-45da-bf4d-2c26e3137d16-kube-api-access-gb9h8\") pod \"openshift-apiserver-operator-796bbdcf4f-m2bq7\" (UID: \"d0cf2421-0839-45da-bf4d-2c26e3137d16\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m2bq7" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402127 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/661771aa-bd40-40fd-8a59-dc3b8156e5c5-client-ca\") pod \"controller-manager-879f6c89f-wdjfz\" (UID: \"661771aa-bd40-40fd-8a59-dc3b8156e5c5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wdjfz" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402145 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ea0a1a12-942c-4ca3-b649-5ad87af478c8-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-sc7xq\" (UID: \"ea0a1a12-942c-4ca3-b649-5ad87af478c8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sc7xq" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402221 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2b720224-3e65-4443-a056-4e8fd226bcea-encryption-config\") pod \"apiserver-7bbb656c7d-bf4gt\" (UID: \"2b720224-3e65-4443-a056-4e8fd226bcea\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bf4gt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402238 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ts95h\" (UniqueName: \"kubernetes.io/projected/2b720224-3e65-4443-a056-4e8fd226bcea-kube-api-access-ts95h\") pod \"apiserver-7bbb656c7d-bf4gt\" (UID: \"2b720224-3e65-4443-a056-4e8fd226bcea\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bf4gt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402256 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/f0515176-8bf5-46a4-a905-7d308f64449e-stats-auth\") pod \"router-default-5444994796-s9lx8\" (UID: \"f0515176-8bf5-46a4-a905-7d308f64449e\") " pod="openshift-ingress/router-default-5444994796-s9lx8" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402272 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/46efdeea-a4d2-4ff5-898d-9ee516105cf4-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vckwz\" (UID: \"46efdeea-a4d2-4ff5-898d-9ee516105cf4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vckwz" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402291 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvvvw\" (UniqueName: \"kubernetes.io/projected/2842cd75-1ff5-4a3f-b7a0-8205388f4a17-kube-api-access-vvvvw\") pod \"route-controller-manager-6576b87f9c-ftgw2\" (UID: \"2842cd75-1ff5-4a3f-b7a0-8205388f4a17\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftgw2" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402309 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9dd57b46-77fc-4b8a-9fc1-a50084165914-console-serving-cert\") pod \"console-f9d7485db-qjxnf\" (UID: \"9dd57b46-77fc-4b8a-9fc1-a50084165914\") " pod="openshift-console/console-f9d7485db-qjxnf" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402326 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/4643a53d-3236-4642-9206-603d20a15129-etcd-ca\") pod \"etcd-operator-b45778765-dckvx\" (UID: \"4643a53d-3236-4642-9206-603d20a15129\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dckvx" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402343 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d0cf2421-0839-45da-bf4d-2c26e3137d16-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-m2bq7\" (UID: \"d0cf2421-0839-45da-bf4d-2c26e3137d16\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m2bq7" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402362 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4643a53d-3236-4642-9206-603d20a15129-etcd-client\") pod \"etcd-operator-b45778765-dckvx\" (UID: \"4643a53d-3236-4642-9206-603d20a15129\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dckvx" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402381 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b4d78aa2-9ff6-404d-9b69-b8da8dfb2f9f-metrics-tls\") pod \"dns-operator-744455d44c-l47x7\" (UID: \"b4d78aa2-9ff6-404d-9b69-b8da8dfb2f9f\") " pod="openshift-dns-operator/dns-operator-744455d44c-l47x7" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402401 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5103e23b-0d63-4e91-8fd9-afd0b755e619-trusted-ca\") pod \"console-operator-58897d9998-jfnms\" (UID: \"5103e23b-0d63-4e91-8fd9-afd0b755e619\") " pod="openshift-console-operator/console-operator-58897d9998-jfnms" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402420 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8eb0c45-6539-478d-96ff-1d43d5d9978b-config\") pod \"machine-approver-56656f9798-p76w8\" (UID: \"e8eb0c45-6539-478d-96ff-1d43d5d9978b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p76w8" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402435 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402453 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/46efdeea-a4d2-4ff5-898d-9ee516105cf4-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vckwz\" (UID: \"46efdeea-a4d2-4ff5-898d-9ee516105cf4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vckwz" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402479 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5b44f815-36aa-42d8-9c3e-fb8d956c05c2-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-9rnrg\" (UID: \"5b44f815-36aa-42d8-9c3e-fb8d956c05c2\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9rnrg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402487 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8h224" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402495 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chhwt\" (UniqueName: \"kubernetes.io/projected/661771aa-bd40-40fd-8a59-dc3b8156e5c5-kube-api-access-chhwt\") pod \"controller-manager-879f6c89f-wdjfz\" (UID: \"661771aa-bd40-40fd-8a59-dc3b8156e5c5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wdjfz" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402513 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5103e23b-0d63-4e91-8fd9-afd0b755e619-config\") pod \"console-operator-58897d9998-jfnms\" (UID: \"5103e23b-0d63-4e91-8fd9-afd0b755e619\") " pod="openshift-console-operator/console-operator-58897d9998-jfnms" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402567 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2b720224-3e65-4443-a056-4e8fd226bcea-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-bf4gt\" (UID: \"2b720224-3e65-4443-a056-4e8fd226bcea\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bf4gt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402582 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8ec02940-3b96-41aa-8624-8338ee544fe1-audit-dir\") pod \"apiserver-76f77b778f-lk4jg\" (UID: \"8ec02940-3b96-41aa-8624-8338ee544fe1\") " pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402599 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f0515176-8bf5-46a4-a905-7d308f64449e-metrics-certs\") pod \"router-default-5444994796-s9lx8\" (UID: \"f0515176-8bf5-46a4-a905-7d308f64449e\") " pod="openshift-ingress/router-default-5444994796-s9lx8" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402617 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2842cd75-1ff5-4a3f-b7a0-8205388f4a17-serving-cert\") pod \"route-controller-manager-6576b87f9c-ftgw2\" (UID: \"2842cd75-1ff5-4a3f-b7a0-8205388f4a17\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftgw2" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402631 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-audit-dir\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402650 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/8ec02940-3b96-41aa-8624-8338ee544fe1-image-import-ca\") pod \"apiserver-76f77b778f-lk4jg\" (UID: \"8ec02940-3b96-41aa-8624-8338ee544fe1\") " pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402664 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0cf2421-0839-45da-bf4d-2c26e3137d16-config\") pod \"openshift-apiserver-operator-796bbdcf4f-m2bq7\" (UID: \"d0cf2421-0839-45da-bf4d-2c26e3137d16\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m2bq7" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402681 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwj4n\" (UniqueName: \"kubernetes.io/projected/b4d78aa2-9ff6-404d-9b69-b8da8dfb2f9f-kube-api-access-vwj4n\") pod \"dns-operator-744455d44c-l47x7\" (UID: \"b4d78aa2-9ff6-404d-9b69-b8da8dfb2f9f\") " pod="openshift-dns-operator/dns-operator-744455d44c-l47x7" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402699 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb0b27d1-5cb0-4a2c-95d7-012bc0a2d071-service-ca-bundle\") pod \"authentication-operator-69f744f599-v676n\" (UID: \"bb0b27d1-5cb0-4a2c-95d7-012bc0a2d071\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v676n" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402715 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/11f26882-b509-4819-9fbb-2c69c062c951-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-lpb68\" (UID: \"11f26882-b509-4819-9fbb-2c69c062c951\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpb68" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402731 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/11f26882-b509-4819-9fbb-2c69c062c951-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-lpb68\" (UID: \"11f26882-b509-4819-9fbb-2c69c062c951\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpb68" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402751 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b720224-3e65-4443-a056-4e8fd226bcea-serving-cert\") pod \"apiserver-7bbb656c7d-bf4gt\" (UID: \"2b720224-3e65-4443-a056-4e8fd226bcea\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bf4gt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402767 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2842cd75-1ff5-4a3f-b7a0-8205388f4a17-config\") pod \"route-controller-manager-6576b87f9c-ftgw2\" (UID: \"2842cd75-1ff5-4a3f-b7a0-8205388f4a17\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftgw2" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.402781 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/661771aa-bd40-40fd-8a59-dc3b8156e5c5-serving-cert\") pod \"controller-manager-879f6c89f-wdjfz\" (UID: \"661771aa-bd40-40fd-8a59-dc3b8156e5c5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wdjfz" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.403024 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.403041 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3508091-c5c6-4dca-9b32-883426154a6e-config\") pod \"machine-api-operator-5694c8668f-zbnls\" (UID: \"d3508091-c5c6-4dca-9b32-883426154a6e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zbnls" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.403056 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8ec02940-3b96-41aa-8624-8338ee544fe1-etcd-serving-ca\") pod \"apiserver-76f77b778f-lk4jg\" (UID: \"8ec02940-3b96-41aa-8624-8338ee544fe1\") " pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.403071 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.403088 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.403105 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d3508091-c5c6-4dca-9b32-883426154a6e-images\") pod \"machine-api-operator-5694c8668f-zbnls\" (UID: \"d3508091-c5c6-4dca-9b32-883426154a6e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zbnls" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.403119 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/4643a53d-3236-4642-9206-603d20a15129-etcd-service-ca\") pod \"etcd-operator-b45778765-dckvx\" (UID: \"4643a53d-3236-4642-9206-603d20a15129\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dckvx" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.403135 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/661771aa-bd40-40fd-8a59-dc3b8156e5c5-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-wdjfz\" (UID: \"661771aa-bd40-40fd-8a59-dc3b8156e5c5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wdjfz" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.403149 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4643a53d-3236-4642-9206-603d20a15129-config\") pod \"etcd-operator-b45778765-dckvx\" (UID: \"4643a53d-3236-4642-9206-603d20a15129\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dckvx" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.403174 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb0b27d1-5cb0-4a2c-95d7-012bc0a2d071-config\") pod \"authentication-operator-69f744f599-v676n\" (UID: \"bb0b27d1-5cb0-4a2c-95d7-012bc0a2d071\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v676n" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.403194 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8ec02940-3b96-41aa-8624-8338ee544fe1-node-pullsecrets\") pod \"apiserver-76f77b778f-lk4jg\" (UID: \"8ec02940-3b96-41aa-8624-8338ee544fe1\") " pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.403220 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f0515176-8bf5-46a4-a905-7d308f64449e-service-ca-bundle\") pod \"router-default-5444994796-s9lx8\" (UID: \"f0515176-8bf5-46a4-a905-7d308f64449e\") " pod="openshift-ingress/router-default-5444994796-s9lx8" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.403275 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.403305 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea0a1a12-942c-4ca3-b649-5ad87af478c8-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-sc7xq\" (UID: \"ea0a1a12-942c-4ca3-b649-5ad87af478c8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sc7xq" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.403331 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/11f26882-b509-4819-9fbb-2c69c062c951-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-lpb68\" (UID: \"11f26882-b509-4819-9fbb-2c69c062c951\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpb68" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.403352 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdvvd\" (UniqueName: \"kubernetes.io/projected/11f26882-b509-4819-9fbb-2c69c062c951-kube-api-access-qdvvd\") pod \"cluster-image-registry-operator-dc59b4c8b-lpb68\" (UID: \"11f26882-b509-4819-9fbb-2c69c062c951\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpb68" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.403401 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ec02940-3b96-41aa-8624-8338ee544fe1-trusted-ca-bundle\") pod \"apiserver-76f77b778f-lk4jg\" (UID: \"8ec02940-3b96-41aa-8624-8338ee544fe1\") " pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.403422 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/f0515176-8bf5-46a4-a905-7d308f64449e-default-certificate\") pod \"router-default-5444994796-s9lx8\" (UID: \"f0515176-8bf5-46a4-a905-7d308f64449e\") " pod="openshift-ingress/router-default-5444994796-s9lx8" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.403450 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9dd57b46-77fc-4b8a-9fc1-a50084165914-console-config\") pod \"console-f9d7485db-qjxnf\" (UID: \"9dd57b46-77fc-4b8a-9fc1-a50084165914\") " pod="openshift-console/console-f9d7485db-qjxnf" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.403471 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.403489 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.403506 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmbkh\" (UniqueName: \"kubernetes.io/projected/f0515176-8bf5-46a4-a905-7d308f64449e-kube-api-access-fmbkh\") pod \"router-default-5444994796-s9lx8\" (UID: \"f0515176-8bf5-46a4-a905-7d308f64449e\") " pod="openshift-ingress/router-default-5444994796-s9lx8" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.403526 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2b720224-3e65-4443-a056-4e8fd226bcea-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-bf4gt\" (UID: \"2b720224-3e65-4443-a056-4e8fd226bcea\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bf4gt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.403571 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9dd57b46-77fc-4b8a-9fc1-a50084165914-service-ca\") pod \"console-f9d7485db-qjxnf\" (UID: \"9dd57b46-77fc-4b8a-9fc1-a50084165914\") " pod="openshift-console/console-f9d7485db-qjxnf" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.403599 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/e8eb0c45-6539-478d-96ff-1d43d5d9978b-machine-approver-tls\") pod \"machine-approver-56656f9798-p76w8\" (UID: \"e8eb0c45-6539-478d-96ff-1d43d5d9978b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p76w8" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.403664 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-6spsg"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.404992 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-6spsg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.406001 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/661771aa-bd40-40fd-8a59-dc3b8156e5c5-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-wdjfz\" (UID: \"661771aa-bd40-40fd-8a59-dc3b8156e5c5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wdjfz" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.406633 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb0b27d1-5cb0-4a2c-95d7-012bc0a2d071-config\") pod \"authentication-operator-69f744f599-v676n\" (UID: \"bb0b27d1-5cb0-4a2c-95d7-012bc0a2d071\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v676n" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.407363 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb0b27d1-5cb0-4a2c-95d7-012bc0a2d071-service-ca-bundle\") pod \"authentication-operator-69f744f599-v676n\" (UID: \"bb0b27d1-5cb0-4a2c-95d7-012bc0a2d071\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v676n" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.407571 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.407673 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/d3508091-c5c6-4dca-9b32-883426154a6e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-zbnls\" (UID: \"d3508091-c5c6-4dca-9b32-883426154a6e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zbnls" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.407766 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9dd57b46-77fc-4b8a-9fc1-a50084165914-oauth-serving-cert\") pod \"console-f9d7485db-qjxnf\" (UID: \"9dd57b46-77fc-4b8a-9fc1-a50084165914\") " pod="openshift-console/console-f9d7485db-qjxnf" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.407842 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfdml\" (UniqueName: \"kubernetes.io/projected/d3508091-c5c6-4dca-9b32-883426154a6e-kube-api-access-zfdml\") pod \"machine-api-operator-5694c8668f-zbnls\" (UID: \"d3508091-c5c6-4dca-9b32-883426154a6e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zbnls" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.403798 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/661771aa-bd40-40fd-8a59-dc3b8156e5c5-config\") pod \"controller-manager-879f6c89f-wdjfz\" (UID: \"661771aa-bd40-40fd-8a59-dc3b8156e5c5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wdjfz" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.414832 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9dd57b46-77fc-4b8a-9fc1-a50084165914-console-config\") pod \"console-f9d7485db-qjxnf\" (UID: \"9dd57b46-77fc-4b8a-9fc1-a50084165914\") " pod="openshift-console/console-f9d7485db-qjxnf" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.415920 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9dd57b46-77fc-4b8a-9fc1-a50084165914-trusted-ca-bundle\") pod \"console-f9d7485db-qjxnf\" (UID: \"9dd57b46-77fc-4b8a-9fc1-a50084165914\") " pod="openshift-console/console-f9d7485db-qjxnf" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.415931 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/661771aa-bd40-40fd-8a59-dc3b8156e5c5-serving-cert\") pod \"controller-manager-879f6c89f-wdjfz\" (UID: \"661771aa-bd40-40fd-8a59-dc3b8156e5c5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wdjfz" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.416132 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ba2ddf1-a9b7-448a-9e48-cdc9e80f3873-config\") pod \"kube-apiserver-operator-766d6c64bb-4942b\" (UID: \"0ba2ddf1-a9b7-448a-9e48-cdc9e80f3873\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4942b" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.416212 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ec02940-3b96-41aa-8624-8338ee544fe1-serving-cert\") pod \"apiserver-76f77b778f-lk4jg\" (UID: \"8ec02940-3b96-41aa-8624-8338ee544fe1\") " pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.416305 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2842cd75-1ff5-4a3f-b7a0-8205388f4a17-config\") pod \"route-controller-manager-6576b87f9c-ftgw2\" (UID: \"2842cd75-1ff5-4a3f-b7a0-8205388f4a17\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftgw2" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.416313 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9dd57b46-77fc-4b8a-9fc1-a50084165914-oauth-serving-cert\") pod \"console-f9d7485db-qjxnf\" (UID: \"9dd57b46-77fc-4b8a-9fc1-a50084165914\") " pod="openshift-console/console-f9d7485db-qjxnf" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.416393 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mt9m\" (UniqueName: \"kubernetes.io/projected/5103e23b-0d63-4e91-8fd9-afd0b755e619-kube-api-access-4mt9m\") pod \"console-operator-58897d9998-jfnms\" (UID: \"5103e23b-0d63-4e91-8fd9-afd0b755e619\") " pod="openshift-console-operator/console-operator-58897d9998-jfnms" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.416026 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9dd57b46-77fc-4b8a-9fc1-a50084165914-console-serving-cert\") pod \"console-f9d7485db-qjxnf\" (UID: \"9dd57b46-77fc-4b8a-9fc1-a50084165914\") " pod="openshift-console/console-f9d7485db-qjxnf" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.416502 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb0b27d1-5cb0-4a2c-95d7-012bc0a2d071-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-v676n\" (UID: \"bb0b27d1-5cb0-4a2c-95d7-012bc0a2d071\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v676n" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.416646 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.417236 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9dd57b46-77fc-4b8a-9fc1-a50084165914-service-ca\") pod \"console-f9d7485db-qjxnf\" (UID: \"9dd57b46-77fc-4b8a-9fc1-a50084165914\") " pod="openshift-console/console-f9d7485db-qjxnf" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.417293 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320665-rkbh5"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.417421 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ec02940-3b96-41aa-8624-8338ee544fe1-config\") pod \"apiserver-76f77b778f-lk4jg\" (UID: \"8ec02940-3b96-41aa-8624-8338ee544fe1\") " pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.417505 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2842cd75-1ff5-4a3f-b7a0-8205388f4a17-client-ca\") pod \"route-controller-manager-6576b87f9c-ftgw2\" (UID: \"2842cd75-1ff5-4a3f-b7a0-8205388f4a17\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftgw2" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.417595 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9dd57b46-77fc-4b8a-9fc1-a50084165914-console-oauth-config\") pod \"console-f9d7485db-qjxnf\" (UID: \"9dd57b46-77fc-4b8a-9fc1-a50084165914\") " pod="openshift-console/console-f9d7485db-qjxnf" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.417640 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpmvq\" (UniqueName: \"kubernetes.io/projected/9dd57b46-77fc-4b8a-9fc1-a50084165914-kube-api-access-cpmvq\") pod \"console-f9d7485db-qjxnf\" (UID: \"9dd57b46-77fc-4b8a-9fc1-a50084165914\") " pod="openshift-console/console-f9d7485db-qjxnf" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.417716 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-audit-policies\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.417814 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5103e23b-0d63-4e91-8fd9-afd0b755e619-serving-cert\") pod \"console-operator-58897d9998-jfnms\" (UID: \"5103e23b-0d63-4e91-8fd9-afd0b755e619\") " pod="openshift-console-operator/console-operator-58897d9998-jfnms" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.417867 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.417906 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8ec02940-3b96-41aa-8624-8338ee544fe1-encryption-config\") pod \"apiserver-76f77b778f-lk4jg\" (UID: \"8ec02940-3b96-41aa-8624-8338ee544fe1\") " pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.418002 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/661771aa-bd40-40fd-8a59-dc3b8156e5c5-client-ca\") pod \"controller-manager-879f6c89f-wdjfz\" (UID: \"661771aa-bd40-40fd-8a59-dc3b8156e5c5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wdjfz" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.418922 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320665-rkbh5" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.418988 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb0b27d1-5cb0-4a2c-95d7-012bc0a2d071-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-v676n\" (UID: \"bb0b27d1-5cb0-4a2c-95d7-012bc0a2d071\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v676n" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.419204 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9sxdn"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.420911 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2842cd75-1ff5-4a3f-b7a0-8205388f4a17-client-ca\") pod \"route-controller-manager-6576b87f9c-ftgw2\" (UID: \"2842cd75-1ff5-4a3f-b7a0-8205388f4a17\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftgw2" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.421486 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.421809 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9sxdn" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.422084 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5b44f815-36aa-42d8-9c3e-fb8d956c05c2-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-9rnrg\" (UID: \"5b44f815-36aa-42d8-9c3e-fb8d956c05c2\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9rnrg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.423832 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9dd57b46-77fc-4b8a-9fc1-a50084165914-console-oauth-config\") pod \"console-f9d7485db-qjxnf\" (UID: \"9dd57b46-77fc-4b8a-9fc1-a50084165914\") " pod="openshift-console/console-f9d7485db-qjxnf" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.427697 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-qjxnf"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.431284 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2842cd75-1ff5-4a3f-b7a0-8205388f4a17-serving-cert\") pod \"route-controller-manager-6576b87f9c-ftgw2\" (UID: \"2842cd75-1ff5-4a3f-b7a0-8205388f4a17\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftgw2" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.431701 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb0b27d1-5cb0-4a2c-95d7-012bc0a2d071-serving-cert\") pod \"authentication-operator-69f744f599-v676n\" (UID: \"bb0b27d1-5cb0-4a2c-95d7-012bc0a2d071\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v676n" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.436137 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-v676n"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.436729 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.437601 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sc7xq"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.440256 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-dckvx"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.441531 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9szj8"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.444191 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-l47x7"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.446080 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9rnrg"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.447446 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-jfnms"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.448753 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z866q"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.449849 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4942b"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.450821 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-czbs2"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.451847 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-jslz5"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.453131 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vckwz"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.454523 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m2bq7"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.454723 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.455798 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-kdwqh"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.456812 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ncblm"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.457878 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-bgjpl"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.458970 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-zbnls"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.459989 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-bf4gt"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.461073 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9sxdn"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.462116 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-snznw"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.463220 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kkmmf"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.464569 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpb68"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.465525 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-8j25v"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.467131 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n2cvx"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.467217 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-8j25v" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.467771 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-56p76"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.468665 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-56p76" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.469260 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s72m9"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.470662 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-6spsg"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.471939 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-lzwls"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.473072 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320665-rkbh5"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.474403 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-9s79z"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.474956 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.475280 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-ptz4x"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.476511 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-dksqg"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.477658 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-lk4jg"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.478790 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-8j25v"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.480004 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-68gcw"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.481101 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-w4vtf"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.482146 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-8h224"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.483223 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-56p76"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.484159 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-zb5zq"] Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.484930 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-zb5zq" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.495982 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.515292 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.519367 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfdml\" (UniqueName: \"kubernetes.io/projected/d3508091-c5c6-4dca-9b32-883426154a6e-kube-api-access-zfdml\") pod \"machine-api-operator-5694c8668f-zbnls\" (UID: \"d3508091-c5c6-4dca-9b32-883426154a6e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zbnls" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.519410 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/713dae78-c3ff-40b4-a018-f801cf0562dd-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-ptz4x\" (UID: \"713dae78-c3ff-40b4-a018-f801cf0562dd\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ptz4x" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.519448 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/94d7d8b4-fc87-48bc-ac61-84e39213d89a-config-volume\") pod \"dns-default-68gcw\" (UID: \"94d7d8b4-fc87-48bc-ac61-84e39213d89a\") " pod="openshift-dns/dns-default-68gcw" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.519526 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ba2ddf1-a9b7-448a-9e48-cdc9e80f3873-config\") pod \"kube-apiserver-operator-766d6c64bb-4942b\" (UID: \"0ba2ddf1-a9b7-448a-9e48-cdc9e80f3873\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4942b" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.519565 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ec02940-3b96-41aa-8624-8338ee544fe1-serving-cert\") pod \"apiserver-76f77b778f-lk4jg\" (UID: \"8ec02940-3b96-41aa-8624-8338ee544fe1\") " pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.519592 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mt9m\" (UniqueName: \"kubernetes.io/projected/5103e23b-0d63-4e91-8fd9-afd0b755e619-kube-api-access-4mt9m\") pod \"console-operator-58897d9998-jfnms\" (UID: \"5103e23b-0d63-4e91-8fd9-afd0b755e619\") " pod="openshift-console-operator/console-operator-58897d9998-jfnms" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.519619 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.519643 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ec02940-3b96-41aa-8624-8338ee544fe1-config\") pod \"apiserver-76f77b778f-lk4jg\" (UID: \"8ec02940-3b96-41aa-8624-8338ee544fe1\") " pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.519666 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-audit-policies\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.519691 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.519721 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkk5g\" (UniqueName: \"kubernetes.io/projected/32d505eb-6035-488f-b154-2699e8af36ed-kube-api-access-qkk5g\") pod \"marketplace-operator-79b997595-snznw\" (UID: \"32d505eb-6035-488f-b154-2699e8af36ed\") " pod="openshift-marketplace/marketplace-operator-79b997595-snznw" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.519750 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmm2z\" (UniqueName: \"kubernetes.io/projected/8ec02940-3b96-41aa-8624-8338ee544fe1-kube-api-access-fmm2z\") pod \"apiserver-76f77b778f-lk4jg\" (UID: \"8ec02940-3b96-41aa-8624-8338ee544fe1\") " pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.519775 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7brf9\" (UniqueName: \"kubernetes.io/projected/e8eb0c45-6539-478d-96ff-1d43d5d9978b-kube-api-access-7brf9\") pod \"machine-approver-56656f9798-p76w8\" (UID: \"e8eb0c45-6539-478d-96ff-1d43d5d9978b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p76w8" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.519798 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvhjm\" (UniqueName: \"kubernetes.io/projected/4643a53d-3236-4642-9206-603d20a15129-kube-api-access-bvhjm\") pod \"etcd-operator-b45778765-dckvx\" (UID: \"4643a53d-3236-4642-9206-603d20a15129\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dckvx" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.519838 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8d0002f3-3315-4c27-bc39-2e5989b67d6b-serving-cert\") pod \"openshift-config-operator-7777fb866f-9s79z\" (UID: \"8d0002f3-3315-4c27-bc39-2e5989b67d6b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9s79z" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.519863 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46efdeea-a4d2-4ff5-898d-9ee516105cf4-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vckwz\" (UID: \"46efdeea-a4d2-4ff5-898d-9ee516105cf4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vckwz" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.519899 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2b720224-3e65-4443-a056-4e8fd226bcea-etcd-client\") pod \"apiserver-7bbb656c7d-bf4gt\" (UID: \"2b720224-3e65-4443-a056-4e8fd226bcea\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bf4gt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.519922 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8ec02940-3b96-41aa-8624-8338ee544fe1-etcd-client\") pod \"apiserver-76f77b778f-lk4jg\" (UID: \"8ec02940-3b96-41aa-8624-8338ee544fe1\") " pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.519944 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2b720224-3e65-4443-a056-4e8fd226bcea-audit-dir\") pod \"apiserver-7bbb656c7d-bf4gt\" (UID: \"2b720224-3e65-4443-a056-4e8fd226bcea\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bf4gt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.519975 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2b720224-3e65-4443-a056-4e8fd226bcea-encryption-config\") pod \"apiserver-7bbb656c7d-bf4gt\" (UID: \"2b720224-3e65-4443-a056-4e8fd226bcea\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bf4gt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.519992 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ts95h\" (UniqueName: \"kubernetes.io/projected/2b720224-3e65-4443-a056-4e8fd226bcea-kube-api-access-ts95h\") pod \"apiserver-7bbb656c7d-bf4gt\" (UID: \"2b720224-3e65-4443-a056-4e8fd226bcea\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bf4gt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.520007 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/f0515176-8bf5-46a4-a905-7d308f64449e-stats-auth\") pod \"router-default-5444994796-s9lx8\" (UID: \"f0515176-8bf5-46a4-a905-7d308f64449e\") " pod="openshift-ingress/router-default-5444994796-s9lx8" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.520022 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/46efdeea-a4d2-4ff5-898d-9ee516105cf4-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vckwz\" (UID: \"46efdeea-a4d2-4ff5-898d-9ee516105cf4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vckwz" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.520038 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5103e23b-0d63-4e91-8fd9-afd0b755e619-trusted-ca\") pod \"console-operator-58897d9998-jfnms\" (UID: \"5103e23b-0d63-4e91-8fd9-afd0b755e619\") " pod="openshift-console-operator/console-operator-58897d9998-jfnms" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.520055 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8eb0c45-6539-478d-96ff-1d43d5d9978b-config\") pod \"machine-approver-56656f9798-p76w8\" (UID: \"e8eb0c45-6539-478d-96ff-1d43d5d9978b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p76w8" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.520070 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b4d78aa2-9ff6-404d-9b69-b8da8dfb2f9f-metrics-tls\") pod \"dns-operator-744455d44c-l47x7\" (UID: \"b4d78aa2-9ff6-404d-9b69-b8da8dfb2f9f\") " pod="openshift-dns-operator/dns-operator-744455d44c-l47x7" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.520096 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5103e23b-0d63-4e91-8fd9-afd0b755e619-config\") pod \"console-operator-58897d9998-jfnms\" (UID: \"5103e23b-0d63-4e91-8fd9-afd0b755e619\") " pod="openshift-console-operator/console-operator-58897d9998-jfnms" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.520113 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.520128 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8ec02940-3b96-41aa-8624-8338ee544fe1-audit-dir\") pod \"apiserver-76f77b778f-lk4jg\" (UID: \"8ec02940-3b96-41aa-8624-8338ee544fe1\") " pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.520143 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-audit-dir\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.520159 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/11f26882-b509-4819-9fbb-2c69c062c951-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-lpb68\" (UID: \"11f26882-b509-4819-9fbb-2c69c062c951\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpb68" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.520178 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d3508091-c5c6-4dca-9b32-883426154a6e-images\") pod \"machine-api-operator-5694c8668f-zbnls\" (UID: \"d3508091-c5c6-4dca-9b32-883426154a6e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zbnls" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.520193 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/4643a53d-3236-4642-9206-603d20a15129-etcd-service-ca\") pod \"etcd-operator-b45778765-dckvx\" (UID: \"4643a53d-3236-4642-9206-603d20a15129\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dckvx" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.520209 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4643a53d-3236-4642-9206-603d20a15129-config\") pod \"etcd-operator-b45778765-dckvx\" (UID: \"4643a53d-3236-4642-9206-603d20a15129\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dckvx" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.520233 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8ec02940-3b96-41aa-8624-8338ee544fe1-node-pullsecrets\") pod \"apiserver-76f77b778f-lk4jg\" (UID: \"8ec02940-3b96-41aa-8624-8338ee544fe1\") " pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.520249 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f0515176-8bf5-46a4-a905-7d308f64449e-service-ca-bundle\") pod \"router-default-5444994796-s9lx8\" (UID: \"f0515176-8bf5-46a4-a905-7d308f64449e\") " pod="openshift-ingress/router-default-5444994796-s9lx8" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.520267 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea0a1a12-942c-4ca3-b649-5ad87af478c8-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-sc7xq\" (UID: \"ea0a1a12-942c-4ca3-b649-5ad87af478c8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sc7xq" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.520286 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.520302 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ec02940-3b96-41aa-8624-8338ee544fe1-trusted-ca-bundle\") pod \"apiserver-76f77b778f-lk4jg\" (UID: \"8ec02940-3b96-41aa-8624-8338ee544fe1\") " pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.520319 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/32d505eb-6035-488f-b154-2699e8af36ed-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-snznw\" (UID: \"32d505eb-6035-488f-b154-2699e8af36ed\") " pod="openshift-marketplace/marketplace-operator-79b997595-snznw" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.520337 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f71cd461-05df-4f5d-bf11-edf4d313996b-config\") pod \"kube-controller-manager-operator-78b949d7b-9szj8\" (UID: \"f71cd461-05df-4f5d-bf11-edf4d313996b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9szj8" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.520358 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.520375 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmbkh\" (UniqueName: \"kubernetes.io/projected/f0515176-8bf5-46a4-a905-7d308f64449e-kube-api-access-fmbkh\") pod \"router-default-5444994796-s9lx8\" (UID: \"f0515176-8bf5-46a4-a905-7d308f64449e\") " pod="openshift-ingress/router-default-5444994796-s9lx8" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.520395 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/e8eb0c45-6539-478d-96ff-1d43d5d9978b-machine-approver-tls\") pod \"machine-approver-56656f9798-p76w8\" (UID: \"e8eb0c45-6539-478d-96ff-1d43d5d9978b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p76w8" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.520427 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2b720224-3e65-4443-a056-4e8fd226bcea-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-bf4gt\" (UID: \"2b720224-3e65-4443-a056-4e8fd226bcea\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bf4gt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.520452 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/69f4620c-5949-4527-aecb-9765f22e653e-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-9sxdn\" (UID: \"69f4620c-5949-4527-aecb-9765f22e653e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9sxdn" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.520475 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/82fad711-fedd-452d-83a5-cf19fbf89815-proxy-tls\") pod \"machine-config-controller-84d6567774-8h224\" (UID: \"82fad711-fedd-452d-83a5-cf19fbf89815\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8h224" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.520495 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26nwz\" (UniqueName: \"kubernetes.io/projected/69f4620c-5949-4527-aecb-9765f22e653e-kube-api-access-26nwz\") pod \"package-server-manager-789f6589d5-9sxdn\" (UID: \"69f4620c-5949-4527-aecb-9765f22e653e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9sxdn" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.520540 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ec02940-3b96-41aa-8624-8338ee544fe1-config\") pod \"apiserver-76f77b778f-lk4jg\" (UID: \"8ec02940-3b96-41aa-8624-8338ee544fe1\") " pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.520682 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-audit-policies\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.520763 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2b720224-3e65-4443-a056-4e8fd226bcea-audit-dir\") pod \"apiserver-7bbb656c7d-bf4gt\" (UID: \"2b720224-3e65-4443-a056-4e8fd226bcea\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bf4gt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.521032 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.521385 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/82fad711-fedd-452d-83a5-cf19fbf89815-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-8h224\" (UID: \"82fad711-fedd-452d-83a5-cf19fbf89815\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8h224" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.521529 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2b720224-3e65-4443-a056-4e8fd226bcea-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-bf4gt\" (UID: \"2b720224-3e65-4443-a056-4e8fd226bcea\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bf4gt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.521629 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.521680 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8ec02940-3b96-41aa-8624-8338ee544fe1-audit-dir\") pod \"apiserver-76f77b778f-lk4jg\" (UID: \"8ec02940-3b96-41aa-8624-8338ee544fe1\") " pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.522115 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5103e23b-0d63-4e91-8fd9-afd0b755e619-trusted-ca\") pod \"console-operator-58897d9998-jfnms\" (UID: \"5103e23b-0d63-4e91-8fd9-afd0b755e619\") " pod="openshift-console-operator/console-operator-58897d9998-jfnms" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.522809 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8eb0c45-6539-478d-96ff-1d43d5d9978b-config\") pod \"machine-approver-56656f9798-p76w8\" (UID: \"e8eb0c45-6539-478d-96ff-1d43d5d9978b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p76w8" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.522858 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ec02940-3b96-41aa-8624-8338ee544fe1-trusted-ca-bundle\") pod \"apiserver-76f77b778f-lk4jg\" (UID: \"8ec02940-3b96-41aa-8624-8338ee544fe1\") " pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.522931 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/4643a53d-3236-4642-9206-603d20a15129-etcd-service-ca\") pod \"etcd-operator-b45778765-dckvx\" (UID: \"4643a53d-3236-4642-9206-603d20a15129\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dckvx" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.522976 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.523512 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/11f26882-b509-4819-9fbb-2c69c062c951-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-lpb68\" (UID: \"11f26882-b509-4819-9fbb-2c69c062c951\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpb68" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.523527 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2b720224-3e65-4443-a056-4e8fd226bcea-encryption-config\") pod \"apiserver-7bbb656c7d-bf4gt\" (UID: \"2b720224-3e65-4443-a056-4e8fd226bcea\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bf4gt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.523572 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea0a1a12-942c-4ca3-b649-5ad87af478c8-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-sc7xq\" (UID: \"ea0a1a12-942c-4ca3-b649-5ad87af478c8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sc7xq" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.523578 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8ec02940-3b96-41aa-8624-8338ee544fe1-node-pullsecrets\") pod \"apiserver-76f77b778f-lk4jg\" (UID: \"8ec02940-3b96-41aa-8624-8338ee544fe1\") " pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.523614 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5103e23b-0d63-4e91-8fd9-afd0b755e619-serving-cert\") pod \"console-operator-58897d9998-jfnms\" (UID: \"5103e23b-0d63-4e91-8fd9-afd0b755e619\") " pod="openshift-console-operator/console-operator-58897d9998-jfnms" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.523638 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2b720224-3e65-4443-a056-4e8fd226bcea-etcd-client\") pod \"apiserver-7bbb656c7d-bf4gt\" (UID: \"2b720224-3e65-4443-a056-4e8fd226bcea\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bf4gt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.523684 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5103e23b-0d63-4e91-8fd9-afd0b755e619-config\") pod \"console-operator-58897d9998-jfnms\" (UID: \"5103e23b-0d63-4e91-8fd9-afd0b755e619\") " pod="openshift-console-operator/console-operator-58897d9998-jfnms" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.523709 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ec02940-3b96-41aa-8624-8338ee544fe1-serving-cert\") pod \"apiserver-76f77b778f-lk4jg\" (UID: \"8ec02940-3b96-41aa-8624-8338ee544fe1\") " pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.523751 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8ec02940-3b96-41aa-8624-8338ee544fe1-encryption-config\") pod \"apiserver-76f77b778f-lk4jg\" (UID: \"8ec02940-3b96-41aa-8624-8338ee544fe1\") " pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.523762 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-audit-dir\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.523805 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2b720224-3e65-4443-a056-4e8fd226bcea-audit-policies\") pod \"apiserver-7bbb656c7d-bf4gt\" (UID: \"2b720224-3e65-4443-a056-4e8fd226bcea\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bf4gt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.523904 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e8eb0c45-6539-478d-96ff-1d43d5d9978b-auth-proxy-config\") pod \"machine-approver-56656f9798-p76w8\" (UID: \"e8eb0c45-6539-478d-96ff-1d43d5d9978b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p76w8" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.523934 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0ba2ddf1-a9b7-448a-9e48-cdc9e80f3873-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-4942b\" (UID: \"0ba2ddf1-a9b7-448a-9e48-cdc9e80f3873\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4942b" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.524011 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rg7s\" (UniqueName: \"kubernetes.io/projected/713dae78-c3ff-40b4-a018-f801cf0562dd-kube-api-access-7rg7s\") pod \"multus-admission-controller-857f4d67dd-ptz4x\" (UID: \"713dae78-c3ff-40b4-a018-f801cf0562dd\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ptz4x" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.524044 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/94d7d8b4-fc87-48bc-ac61-84e39213d89a-metrics-tls\") pod \"dns-default-68gcw\" (UID: \"94d7d8b4-fc87-48bc-ac61-84e39213d89a\") " pod="openshift-dns/dns-default-68gcw" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.524062 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4643a53d-3236-4642-9206-603d20a15129-config\") pod \"etcd-operator-b45778765-dckvx\" (UID: \"4643a53d-3236-4642-9206-603d20a15129\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dckvx" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.524139 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0ba2ddf1-a9b7-448a-9e48-cdc9e80f3873-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-4942b\" (UID: \"0ba2ddf1-a9b7-448a-9e48-cdc9e80f3873\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4942b" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.524140 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8d0002f3-3315-4c27-bc39-2e5989b67d6b-serving-cert\") pod \"openshift-config-operator-7777fb866f-9s79z\" (UID: \"8d0002f3-3315-4c27-bc39-2e5989b67d6b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9s79z" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.524168 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8dpt\" (UniqueName: \"kubernetes.io/projected/ea0a1a12-942c-4ca3-b649-5ad87af478c8-kube-api-access-d8dpt\") pod \"openshift-controller-manager-operator-756b6f6bc6-sc7xq\" (UID: \"ea0a1a12-942c-4ca3-b649-5ad87af478c8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sc7xq" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.524338 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4643a53d-3236-4642-9206-603d20a15129-serving-cert\") pod \"etcd-operator-b45778765-dckvx\" (UID: \"4643a53d-3236-4642-9206-603d20a15129\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dckvx" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.524399 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f71cd461-05df-4f5d-bf11-edf4d313996b-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-9szj8\" (UID: \"f71cd461-05df-4f5d-bf11-edf4d313996b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9szj8" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.524709 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2b720224-3e65-4443-a056-4e8fd226bcea-audit-policies\") pod \"apiserver-7bbb656c7d-bf4gt\" (UID: \"2b720224-3e65-4443-a056-4e8fd226bcea\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bf4gt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.524730 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d3508091-c5c6-4dca-9b32-883426154a6e-images\") pod \"machine-api-operator-5694c8668f-zbnls\" (UID: \"d3508091-c5c6-4dca-9b32-883426154a6e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zbnls" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.524769 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.524820 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8d0002f3-3315-4c27-bc39-2e5989b67d6b-available-featuregates\") pod \"openshift-config-operator-7777fb866f-9s79z\" (UID: \"8d0002f3-3315-4c27-bc39-2e5989b67d6b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9s79z" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.524954 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e8eb0c45-6539-478d-96ff-1d43d5d9978b-auth-proxy-config\") pod \"machine-approver-56656f9798-p76w8\" (UID: \"e8eb0c45-6539-478d-96ff-1d43d5d9978b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p76w8" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.525040 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/8ec02940-3b96-41aa-8624-8338ee544fe1-audit\") pod \"apiserver-76f77b778f-lk4jg\" (UID: \"8ec02940-3b96-41aa-8624-8338ee544fe1\") " pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.525096 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8ec02940-3b96-41aa-8624-8338ee544fe1-etcd-client\") pod \"apiserver-76f77b778f-lk4jg\" (UID: \"8ec02940-3b96-41aa-8624-8338ee544fe1\") " pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.525179 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8d0002f3-3315-4c27-bc39-2e5989b67d6b-available-featuregates\") pod \"openshift-config-operator-7777fb866f-9s79z\" (UID: \"8d0002f3-3315-4c27-bc39-2e5989b67d6b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9s79z" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.525201 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmrwg\" (UniqueName: \"kubernetes.io/projected/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-kube-api-access-qmrwg\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.525235 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdp5r\" (UniqueName: \"kubernetes.io/projected/8d0002f3-3315-4c27-bc39-2e5989b67d6b-kube-api-access-jdp5r\") pod \"openshift-config-operator-7777fb866f-9s79z\" (UID: \"8d0002f3-3315-4c27-bc39-2e5989b67d6b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9s79z" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.525310 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gb9h8\" (UniqueName: \"kubernetes.io/projected/d0cf2421-0839-45da-bf4d-2c26e3137d16-kube-api-access-gb9h8\") pod \"openshift-apiserver-operator-796bbdcf4f-m2bq7\" (UID: \"d0cf2421-0839-45da-bf4d-2c26e3137d16\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m2bq7" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.525342 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ea0a1a12-942c-4ca3-b649-5ad87af478c8-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-sc7xq\" (UID: \"ea0a1a12-942c-4ca3-b649-5ad87af478c8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sc7xq" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.525368 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/4643a53d-3236-4642-9206-603d20a15129-etcd-ca\") pod \"etcd-operator-b45778765-dckvx\" (UID: \"4643a53d-3236-4642-9206-603d20a15129\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dckvx" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.525396 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d0cf2421-0839-45da-bf4d-2c26e3137d16-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-m2bq7\" (UID: \"d0cf2421-0839-45da-bf4d-2c26e3137d16\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m2bq7" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.525437 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4643a53d-3236-4642-9206-603d20a15129-etcd-client\") pod \"etcd-operator-b45778765-dckvx\" (UID: \"4643a53d-3236-4642-9206-603d20a15129\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dckvx" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.525460 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/46efdeea-a4d2-4ff5-898d-9ee516105cf4-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vckwz\" (UID: \"46efdeea-a4d2-4ff5-898d-9ee516105cf4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vckwz" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.525495 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2b720224-3e65-4443-a056-4e8fd226bcea-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-bf4gt\" (UID: \"2b720224-3e65-4443-a056-4e8fd226bcea\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bf4gt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.525520 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f0515176-8bf5-46a4-a905-7d308f64449e-metrics-certs\") pod \"router-default-5444994796-s9lx8\" (UID: \"f0515176-8bf5-46a4-a905-7d308f64449e\") " pod="openshift-ingress/router-default-5444994796-s9lx8" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.525544 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/8ec02940-3b96-41aa-8624-8338ee544fe1-image-import-ca\") pod \"apiserver-76f77b778f-lk4jg\" (UID: \"8ec02940-3b96-41aa-8624-8338ee544fe1\") " pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.525604 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0cf2421-0839-45da-bf4d-2c26e3137d16-config\") pod \"openshift-apiserver-operator-796bbdcf4f-m2bq7\" (UID: \"d0cf2421-0839-45da-bf4d-2c26e3137d16\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m2bq7" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.525629 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwj4n\" (UniqueName: \"kubernetes.io/projected/b4d78aa2-9ff6-404d-9b69-b8da8dfb2f9f-kube-api-access-vwj4n\") pod \"dns-operator-744455d44c-l47x7\" (UID: \"b4d78aa2-9ff6-404d-9b69-b8da8dfb2f9f\") " pod="openshift-dns-operator/dns-operator-744455d44c-l47x7" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.525656 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/32d505eb-6035-488f-b154-2699e8af36ed-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-snznw\" (UID: \"32d505eb-6035-488f-b154-2699e8af36ed\") " pod="openshift-marketplace/marketplace-operator-79b997595-snznw" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.525661 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/8ec02940-3b96-41aa-8624-8338ee544fe1-audit\") pod \"apiserver-76f77b778f-lk4jg\" (UID: \"8ec02940-3b96-41aa-8624-8338ee544fe1\") " pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.525685 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.525712 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3508091-c5c6-4dca-9b32-883426154a6e-config\") pod \"machine-api-operator-5694c8668f-zbnls\" (UID: \"d3508091-c5c6-4dca-9b32-883426154a6e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zbnls" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.525736 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/11f26882-b509-4819-9fbb-2c69c062c951-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-lpb68\" (UID: \"11f26882-b509-4819-9fbb-2c69c062c951\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpb68" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.525764 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b720224-3e65-4443-a056-4e8fd226bcea-serving-cert\") pod \"apiserver-7bbb656c7d-bf4gt\" (UID: \"2b720224-3e65-4443-a056-4e8fd226bcea\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bf4gt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.525790 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8ec02940-3b96-41aa-8624-8338ee544fe1-etcd-serving-ca\") pod \"apiserver-76f77b778f-lk4jg\" (UID: \"8ec02940-3b96-41aa-8624-8338ee544fe1\") " pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.525812 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.525835 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.525888 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnq5m\" (UniqueName: \"kubernetes.io/projected/82fad711-fedd-452d-83a5-cf19fbf89815-kube-api-access-qnq5m\") pod \"machine-config-controller-84d6567774-8h224\" (UID: \"82fad711-fedd-452d-83a5-cf19fbf89815\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8h224" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.525917 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/11f26882-b509-4819-9fbb-2c69c062c951-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-lpb68\" (UID: \"11f26882-b509-4819-9fbb-2c69c062c951\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpb68" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.525957 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdvvd\" (UniqueName: \"kubernetes.io/projected/11f26882-b509-4819-9fbb-2c69c062c951-kube-api-access-qdvvd\") pod \"cluster-image-registry-operator-dc59b4c8b-lpb68\" (UID: \"11f26882-b509-4819-9fbb-2c69c062c951\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpb68" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.525986 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/f0515176-8bf5-46a4-a905-7d308f64449e-default-certificate\") pod \"router-default-5444994796-s9lx8\" (UID: \"f0515176-8bf5-46a4-a905-7d308f64449e\") " pod="openshift-ingress/router-default-5444994796-s9lx8" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.526005 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bx8vf\" (UniqueName: \"kubernetes.io/projected/94d7d8b4-fc87-48bc-ac61-84e39213d89a-kube-api-access-bx8vf\") pod \"dns-default-68gcw\" (UID: \"94d7d8b4-fc87-48bc-ac61-84e39213d89a\") " pod="openshift-dns/dns-default-68gcw" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.526011 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/4643a53d-3236-4642-9206-603d20a15129-etcd-ca\") pod \"etcd-operator-b45778765-dckvx\" (UID: \"4643a53d-3236-4642-9206-603d20a15129\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dckvx" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.526024 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.526043 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f71cd461-05df-4f5d-bf11-edf4d313996b-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-9szj8\" (UID: \"f71cd461-05df-4f5d-bf11-edf4d313996b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9szj8" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.526063 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.526082 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/d3508091-c5c6-4dca-9b32-883426154a6e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-zbnls\" (UID: \"d3508091-c5c6-4dca-9b32-883426154a6e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zbnls" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.526469 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2b720224-3e65-4443-a056-4e8fd226bcea-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-bf4gt\" (UID: \"2b720224-3e65-4443-a056-4e8fd226bcea\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bf4gt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.526470 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8ec02940-3b96-41aa-8624-8338ee544fe1-encryption-config\") pod \"apiserver-76f77b778f-lk4jg\" (UID: \"8ec02940-3b96-41aa-8624-8338ee544fe1\") " pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.526605 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.527030 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.527332 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5103e23b-0d63-4e91-8fd9-afd0b755e619-serving-cert\") pod \"console-operator-58897d9998-jfnms\" (UID: \"5103e23b-0d63-4e91-8fd9-afd0b755e619\") " pod="openshift-console-operator/console-operator-58897d9998-jfnms" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.527613 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.527799 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4643a53d-3236-4642-9206-603d20a15129-serving-cert\") pod \"etcd-operator-b45778765-dckvx\" (UID: \"4643a53d-3236-4642-9206-603d20a15129\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dckvx" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.527841 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3508091-c5c6-4dca-9b32-883426154a6e-config\") pod \"machine-api-operator-5694c8668f-zbnls\" (UID: \"d3508091-c5c6-4dca-9b32-883426154a6e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zbnls" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.527859 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.528177 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/8ec02940-3b96-41aa-8624-8338ee544fe1-image-import-ca\") pod \"apiserver-76f77b778f-lk4jg\" (UID: \"8ec02940-3b96-41aa-8624-8338ee544fe1\") " pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.528386 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8ec02940-3b96-41aa-8624-8338ee544fe1-etcd-serving-ca\") pod \"apiserver-76f77b778f-lk4jg\" (UID: \"8ec02940-3b96-41aa-8624-8338ee544fe1\") " pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.528834 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4643a53d-3236-4642-9206-603d20a15129-etcd-client\") pod \"etcd-operator-b45778765-dckvx\" (UID: \"4643a53d-3236-4642-9206-603d20a15129\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dckvx" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.529503 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.529583 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ea0a1a12-942c-4ca3-b649-5ad87af478c8-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-sc7xq\" (UID: \"ea0a1a12-942c-4ca3-b649-5ad87af478c8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sc7xq" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.529573 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/d3508091-c5c6-4dca-9b32-883426154a6e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-zbnls\" (UID: \"d3508091-c5c6-4dca-9b32-883426154a6e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zbnls" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.529663 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.529686 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.529985 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/e8eb0c45-6539-478d-96ff-1d43d5d9978b-machine-approver-tls\") pod \"machine-approver-56656f9798-p76w8\" (UID: \"e8eb0c45-6539-478d-96ff-1d43d5d9978b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p76w8" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.531632 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0cf2421-0839-45da-bf4d-2c26e3137d16-config\") pod \"openshift-apiserver-operator-796bbdcf4f-m2bq7\" (UID: \"d0cf2421-0839-45da-bf4d-2c26e3137d16\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m2bq7" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.531752 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/11f26882-b509-4819-9fbb-2c69c062c951-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-lpb68\" (UID: \"11f26882-b509-4819-9fbb-2c69c062c951\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpb68" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.532159 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.532684 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b720224-3e65-4443-a056-4e8fd226bcea-serving-cert\") pod \"apiserver-7bbb656c7d-bf4gt\" (UID: \"2b720224-3e65-4443-a056-4e8fd226bcea\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bf4gt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.533194 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d0cf2421-0839-45da-bf4d-2c26e3137d16-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-m2bq7\" (UID: \"d0cf2421-0839-45da-bf4d-2c26e3137d16\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m2bq7" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.537487 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.555767 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.566948 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b4d78aa2-9ff6-404d-9b69-b8da8dfb2f9f-metrics-tls\") pod \"dns-operator-744455d44c-l47x7\" (UID: \"b4d78aa2-9ff6-404d-9b69-b8da8dfb2f9f\") " pod="openshift-dns-operator/dns-operator-744455d44c-l47x7" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.574423 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.595199 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.614986 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.622006 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f0515176-8bf5-46a4-a905-7d308f64449e-service-ca-bundle\") pod \"router-default-5444994796-s9lx8\" (UID: \"f0515176-8bf5-46a4-a905-7d308f64449e\") " pod="openshift-ingress/router-default-5444994796-s9lx8" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.626893 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f71cd461-05df-4f5d-bf11-edf4d313996b-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-9szj8\" (UID: \"f71cd461-05df-4f5d-bf11-edf4d313996b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9szj8" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.626988 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/32d505eb-6035-488f-b154-2699e8af36ed-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-snznw\" (UID: \"32d505eb-6035-488f-b154-2699e8af36ed\") " pod="openshift-marketplace/marketplace-operator-79b997595-snznw" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.627008 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnq5m\" (UniqueName: \"kubernetes.io/projected/82fad711-fedd-452d-83a5-cf19fbf89815-kube-api-access-qnq5m\") pod \"machine-config-controller-84d6567774-8h224\" (UID: \"82fad711-fedd-452d-83a5-cf19fbf89815\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8h224" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.627050 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bx8vf\" (UniqueName: \"kubernetes.io/projected/94d7d8b4-fc87-48bc-ac61-84e39213d89a-kube-api-access-bx8vf\") pod \"dns-default-68gcw\" (UID: \"94d7d8b4-fc87-48bc-ac61-84e39213d89a\") " pod="openshift-dns/dns-default-68gcw" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.627070 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f71cd461-05df-4f5d-bf11-edf4d313996b-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-9szj8\" (UID: \"f71cd461-05df-4f5d-bf11-edf4d313996b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9szj8" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.627087 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/713dae78-c3ff-40b4-a018-f801cf0562dd-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-ptz4x\" (UID: \"713dae78-c3ff-40b4-a018-f801cf0562dd\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ptz4x" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.627109 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/94d7d8b4-fc87-48bc-ac61-84e39213d89a-config-volume\") pod \"dns-default-68gcw\" (UID: \"94d7d8b4-fc87-48bc-ac61-84e39213d89a\") " pod="openshift-dns/dns-default-68gcw" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.627137 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkk5g\" (UniqueName: \"kubernetes.io/projected/32d505eb-6035-488f-b154-2699e8af36ed-kube-api-access-qkk5g\") pod \"marketplace-operator-79b997595-snznw\" (UID: \"32d505eb-6035-488f-b154-2699e8af36ed\") " pod="openshift-marketplace/marketplace-operator-79b997595-snznw" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.627226 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/32d505eb-6035-488f-b154-2699e8af36ed-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-snznw\" (UID: \"32d505eb-6035-488f-b154-2699e8af36ed\") " pod="openshift-marketplace/marketplace-operator-79b997595-snznw" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.627244 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f71cd461-05df-4f5d-bf11-edf4d313996b-config\") pod \"kube-controller-manager-operator-78b949d7b-9szj8\" (UID: \"f71cd461-05df-4f5d-bf11-edf4d313996b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9szj8" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.627265 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/69f4620c-5949-4527-aecb-9765f22e653e-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-9sxdn\" (UID: \"69f4620c-5949-4527-aecb-9765f22e653e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9sxdn" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.627283 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/82fad711-fedd-452d-83a5-cf19fbf89815-proxy-tls\") pod \"machine-config-controller-84d6567774-8h224\" (UID: \"82fad711-fedd-452d-83a5-cf19fbf89815\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8h224" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.627300 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26nwz\" (UniqueName: \"kubernetes.io/projected/69f4620c-5949-4527-aecb-9765f22e653e-kube-api-access-26nwz\") pod \"package-server-manager-789f6589d5-9sxdn\" (UID: \"69f4620c-5949-4527-aecb-9765f22e653e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9sxdn" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.627328 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/82fad711-fedd-452d-83a5-cf19fbf89815-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-8h224\" (UID: \"82fad711-fedd-452d-83a5-cf19fbf89815\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8h224" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.627357 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rg7s\" (UniqueName: \"kubernetes.io/projected/713dae78-c3ff-40b4-a018-f801cf0562dd-kube-api-access-7rg7s\") pod \"multus-admission-controller-857f4d67dd-ptz4x\" (UID: \"713dae78-c3ff-40b4-a018-f801cf0562dd\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ptz4x" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.627374 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/94d7d8b4-fc87-48bc-ac61-84e39213d89a-metrics-tls\") pod \"dns-default-68gcw\" (UID: \"94d7d8b4-fc87-48bc-ac61-84e39213d89a\") " pod="openshift-dns/dns-default-68gcw" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.628885 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/82fad711-fedd-452d-83a5-cf19fbf89815-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-8h224\" (UID: \"82fad711-fedd-452d-83a5-cf19fbf89815\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8h224" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.635090 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.644696 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/f0515176-8bf5-46a4-a905-7d308f64449e-stats-auth\") pod \"router-default-5444994796-s9lx8\" (UID: \"f0515176-8bf5-46a4-a905-7d308f64449e\") " pod="openshift-ingress/router-default-5444994796-s9lx8" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.655305 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.676503 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.680618 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/f0515176-8bf5-46a4-a905-7d308f64449e-default-certificate\") pod \"router-default-5444994796-s9lx8\" (UID: \"f0515176-8bf5-46a4-a905-7d308f64449e\") " pod="openshift-ingress/router-default-5444994796-s9lx8" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.695637 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.698600 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f0515176-8bf5-46a4-a905-7d308f64449e-metrics-certs\") pod \"router-default-5444994796-s9lx8\" (UID: \"f0515176-8bf5-46a4-a905-7d308f64449e\") " pod="openshift-ingress/router-default-5444994796-s9lx8" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.715448 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.736841 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.774480 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.784406 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/46efdeea-a4d2-4ff5-898d-9ee516105cf4-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vckwz\" (UID: \"46efdeea-a4d2-4ff5-898d-9ee516105cf4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vckwz" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.794657 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.801349 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46efdeea-a4d2-4ff5-898d-9ee516105cf4-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vckwz\" (UID: \"46efdeea-a4d2-4ff5-898d-9ee516105cf4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vckwz" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.815453 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.835010 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.855395 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.875757 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.896064 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.908825 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0ba2ddf1-a9b7-448a-9e48-cdc9e80f3873-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-4942b\" (UID: \"0ba2ddf1-a9b7-448a-9e48-cdc9e80f3873\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4942b" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.915426 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.921166 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ba2ddf1-a9b7-448a-9e48-cdc9e80f3873-config\") pod \"kube-apiserver-operator-766d6c64bb-4942b\" (UID: \"0ba2ddf1-a9b7-448a-9e48-cdc9e80f3873\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4942b" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.936632 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.956318 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.975119 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Sep 30 13:58:38 crc kubenswrapper[4840]: I0930 13:58:38.995449 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.001043 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f71cd461-05df-4f5d-bf11-edf4d313996b-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-9szj8\" (UID: \"f71cd461-05df-4f5d-bf11-edf4d313996b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9szj8" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.016809 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.041213 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.054833 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.058498 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f71cd461-05df-4f5d-bf11-edf4d313996b-config\") pod \"kube-controller-manager-operator-78b949d7b-9szj8\" (UID: \"f71cd461-05df-4f5d-bf11-edf4d313996b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9szj8" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.075524 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.096052 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.115377 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.115574 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.115628 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.115577 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.135321 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.155386 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.175340 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.182344 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/713dae78-c3ff-40b4-a018-f801cf0562dd-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-ptz4x\" (UID: \"713dae78-c3ff-40b4-a018-f801cf0562dd\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ptz4x" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.195705 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.215620 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.234923 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.255960 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.275842 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.295625 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.314701 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.335066 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.355315 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.374978 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.393702 4840 request.go:700] Waited for 1.00766113s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/secrets?fieldSelector=metadata.name%3Dolm-operator-serviceaccount-dockercfg-rq7zk&limit=500&resourceVersion=0 Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.396302 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.416011 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.436397 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.456064 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.475492 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.494998 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.516037 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.536199 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.556214 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.582075 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.595224 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.601158 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/32d505eb-6035-488f-b154-2699e8af36ed-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-snznw\" (UID: \"32d505eb-6035-488f-b154-2699e8af36ed\") " pod="openshift-marketplace/marketplace-operator-79b997595-snznw" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.617314 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Sep 30 13:58:39 crc kubenswrapper[4840]: E0930 13:58:39.627631 4840 configmap.go:193] Couldn't get configMap openshift-dns/dns-default: failed to sync configmap cache: timed out waiting for the condition Sep 30 13:58:39 crc kubenswrapper[4840]: E0930 13:58:39.627640 4840 secret.go:188] Couldn't get secret openshift-machine-config-operator/mcc-proxy-tls: failed to sync secret cache: timed out waiting for the condition Sep 30 13:58:39 crc kubenswrapper[4840]: E0930 13:58:39.627661 4840 configmap.go:193] Couldn't get configMap openshift-marketplace/marketplace-trusted-ca: failed to sync configmap cache: timed out waiting for the condition Sep 30 13:58:39 crc kubenswrapper[4840]: E0930 13:58:39.627682 4840 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/package-server-manager-serving-cert: failed to sync secret cache: timed out waiting for the condition Sep 30 13:58:39 crc kubenswrapper[4840]: E0930 13:58:39.627724 4840 secret.go:188] Couldn't get secret openshift-dns/dns-default-metrics-tls: failed to sync secret cache: timed out waiting for the condition Sep 30 13:58:39 crc kubenswrapper[4840]: E0930 13:58:39.628001 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/94d7d8b4-fc87-48bc-ac61-84e39213d89a-config-volume podName:94d7d8b4-fc87-48bc-ac61-84e39213d89a nodeName:}" failed. No retries permitted until 2025-09-30 13:58:40.127965091 +0000 UTC m=+148.757051524 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/94d7d8b4-fc87-48bc-ac61-84e39213d89a-config-volume") pod "dns-default-68gcw" (UID: "94d7d8b4-fc87-48bc-ac61-84e39213d89a") : failed to sync configmap cache: timed out waiting for the condition Sep 30 13:58:39 crc kubenswrapper[4840]: E0930 13:58:39.628313 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82fad711-fedd-452d-83a5-cf19fbf89815-proxy-tls podName:82fad711-fedd-452d-83a5-cf19fbf89815 nodeName:}" failed. No retries permitted until 2025-09-30 13:58:40.128272589 +0000 UTC m=+148.757359202 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-tls" (UniqueName: "kubernetes.io/secret/82fad711-fedd-452d-83a5-cf19fbf89815-proxy-tls") pod "machine-config-controller-84d6567774-8h224" (UID: "82fad711-fedd-452d-83a5-cf19fbf89815") : failed to sync secret cache: timed out waiting for the condition Sep 30 13:58:39 crc kubenswrapper[4840]: E0930 13:58:39.628344 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/32d505eb-6035-488f-b154-2699e8af36ed-marketplace-trusted-ca podName:32d505eb-6035-488f-b154-2699e8af36ed nodeName:}" failed. No retries permitted until 2025-09-30 13:58:40.128335471 +0000 UTC m=+148.757421894 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-trusted-ca" (UniqueName: "kubernetes.io/configmap/32d505eb-6035-488f-b154-2699e8af36ed-marketplace-trusted-ca") pod "marketplace-operator-79b997595-snznw" (UID: "32d505eb-6035-488f-b154-2699e8af36ed") : failed to sync configmap cache: timed out waiting for the condition Sep 30 13:58:39 crc kubenswrapper[4840]: E0930 13:58:39.628370 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/69f4620c-5949-4527-aecb-9765f22e653e-package-server-manager-serving-cert podName:69f4620c-5949-4527-aecb-9765f22e653e nodeName:}" failed. No retries permitted until 2025-09-30 13:58:40.128360972 +0000 UTC m=+148.757447615 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "package-server-manager-serving-cert" (UniqueName: "kubernetes.io/secret/69f4620c-5949-4527-aecb-9765f22e653e-package-server-manager-serving-cert") pod "package-server-manager-789f6589d5-9sxdn" (UID: "69f4620c-5949-4527-aecb-9765f22e653e") : failed to sync secret cache: timed out waiting for the condition Sep 30 13:58:39 crc kubenswrapper[4840]: E0930 13:58:39.628396 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/94d7d8b4-fc87-48bc-ac61-84e39213d89a-metrics-tls podName:94d7d8b4-fc87-48bc-ac61-84e39213d89a nodeName:}" failed. No retries permitted until 2025-09-30 13:58:40.128389482 +0000 UTC m=+148.757476135 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/94d7d8b4-fc87-48bc-ac61-84e39213d89a-metrics-tls") pod "dns-default-68gcw" (UID: "94d7d8b4-fc87-48bc-ac61-84e39213d89a") : failed to sync secret cache: timed out waiting for the condition Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.642505 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.656206 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.676101 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.696270 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.715196 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.736107 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.755581 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.775171 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.796276 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.815867 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.835223 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.855940 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.874900 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.895401 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.929640 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2z4f\" (UniqueName: \"kubernetes.io/projected/bb0b27d1-5cb0-4a2c-95d7-012bc0a2d071-kube-api-access-v2z4f\") pod \"authentication-operator-69f744f599-v676n\" (UID: \"bb0b27d1-5cb0-4a2c-95d7-012bc0a2d071\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v676n" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.935604 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.955933 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.975039 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Sep 30 13:58:39 crc kubenswrapper[4840]: I0930 13:58:39.995585 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.015684 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.054104 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvvvw\" (UniqueName: \"kubernetes.io/projected/2842cd75-1ff5-4a3f-b7a0-8205388f4a17-kube-api-access-vvvvw\") pod \"route-controller-manager-6576b87f9c-ftgw2\" (UID: \"2842cd75-1ff5-4a3f-b7a0-8205388f4a17\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftgw2" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.065895 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftgw2" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.071688 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chhwt\" (UniqueName: \"kubernetes.io/projected/661771aa-bd40-40fd-8a59-dc3b8156e5c5-kube-api-access-chhwt\") pod \"controller-manager-879f6c89f-wdjfz\" (UID: \"661771aa-bd40-40fd-8a59-dc3b8156e5c5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wdjfz" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.090581 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmp25\" (UniqueName: \"kubernetes.io/projected/dde0d17e-6bd0-4acb-b8b8-a8b0bc988a6e-kube-api-access-mmp25\") pod \"downloads-7954f5f757-kdwqh\" (UID: \"dde0d17e-6bd0-4acb-b8b8-a8b0bc988a6e\") " pod="openshift-console/downloads-7954f5f757-kdwqh" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.114670 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-v676n" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.114681 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qr4f5\" (UniqueName: \"kubernetes.io/projected/5b44f815-36aa-42d8-9c3e-fb8d956c05c2-kube-api-access-qr4f5\") pod \"cluster-samples-operator-665b6dd947-9rnrg\" (UID: \"5b44f815-36aa-42d8-9c3e-fb8d956c05c2\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9rnrg" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.133012 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpmvq\" (UniqueName: \"kubernetes.io/projected/9dd57b46-77fc-4b8a-9fc1-a50084165914-kube-api-access-cpmvq\") pod \"console-f9d7485db-qjxnf\" (UID: \"9dd57b46-77fc-4b8a-9fc1-a50084165914\") " pod="openshift-console/console-f9d7485db-qjxnf" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.133504 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-kdwqh" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.135257 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.148527 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/94d7d8b4-fc87-48bc-ac61-84e39213d89a-config-volume\") pod \"dns-default-68gcw\" (UID: \"94d7d8b4-fc87-48bc-ac61-84e39213d89a\") " pod="openshift-dns/dns-default-68gcw" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.149741 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/69f4620c-5949-4527-aecb-9765f22e653e-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-9sxdn\" (UID: \"69f4620c-5949-4527-aecb-9765f22e653e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9sxdn" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.149812 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/82fad711-fedd-452d-83a5-cf19fbf89815-proxy-tls\") pod \"machine-config-controller-84d6567774-8h224\" (UID: \"82fad711-fedd-452d-83a5-cf19fbf89815\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8h224" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.149886 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/94d7d8b4-fc87-48bc-ac61-84e39213d89a-metrics-tls\") pod \"dns-default-68gcw\" (UID: \"94d7d8b4-fc87-48bc-ac61-84e39213d89a\") " pod="openshift-dns/dns-default-68gcw" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.150029 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/32d505eb-6035-488f-b154-2699e8af36ed-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-snznw\" (UID: \"32d505eb-6035-488f-b154-2699e8af36ed\") " pod="openshift-marketplace/marketplace-operator-79b997595-snznw" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.151601 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/32d505eb-6035-488f-b154-2699e8af36ed-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-snznw\" (UID: \"32d505eb-6035-488f-b154-2699e8af36ed\") " pod="openshift-marketplace/marketplace-operator-79b997595-snznw" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.153948 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/82fad711-fedd-452d-83a5-cf19fbf89815-proxy-tls\") pod \"machine-config-controller-84d6567774-8h224\" (UID: \"82fad711-fedd-452d-83a5-cf19fbf89815\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8h224" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.155294 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.176042 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.184226 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/69f4620c-5949-4527-aecb-9765f22e653e-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-9sxdn\" (UID: \"69f4620c-5949-4527-aecb-9765f22e653e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9sxdn" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.215541 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.235861 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.255877 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.275524 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.290174 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/94d7d8b4-fc87-48bc-ac61-84e39213d89a-config-volume\") pod \"dns-default-68gcw\" (UID: \"94d7d8b4-fc87-48bc-ac61-84e39213d89a\") " pod="openshift-dns/dns-default-68gcw" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.290196 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/94d7d8b4-fc87-48bc-ac61-84e39213d89a-metrics-tls\") pod \"dns-default-68gcw\" (UID: \"94d7d8b4-fc87-48bc-ac61-84e39213d89a\") " pod="openshift-dns/dns-default-68gcw" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.295727 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.316110 4840 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.332126 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-wdjfz" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.335235 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.345088 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-qjxnf" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.356883 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.375455 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.392010 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9rnrg" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.395389 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.416831 4840 request.go:700] Waited for 1.896840477s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-api/serviceaccounts/machine-api-operator/token Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.441866 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfdml\" (UniqueName: \"kubernetes.io/projected/d3508091-c5c6-4dca-9b32-883426154a6e-kube-api-access-zfdml\") pod \"machine-api-operator-5694c8668f-zbnls\" (UID: \"d3508091-c5c6-4dca-9b32-883426154a6e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zbnls" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.460494 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mt9m\" (UniqueName: \"kubernetes.io/projected/5103e23b-0d63-4e91-8fd9-afd0b755e619-kube-api-access-4mt9m\") pod \"console-operator-58897d9998-jfnms\" (UID: \"5103e23b-0d63-4e91-8fd9-afd0b755e619\") " pod="openshift-console-operator/console-operator-58897d9998-jfnms" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.473533 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvhjm\" (UniqueName: \"kubernetes.io/projected/4643a53d-3236-4642-9206-603d20a15129-kube-api-access-bvhjm\") pod \"etcd-operator-b45778765-dckvx\" (UID: \"4643a53d-3236-4642-9206-603d20a15129\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dckvx" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.482057 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-zbnls" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.489022 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmm2z\" (UniqueName: \"kubernetes.io/projected/8ec02940-3b96-41aa-8624-8338ee544fe1-kube-api-access-fmm2z\") pod \"apiserver-76f77b778f-lk4jg\" (UID: \"8ec02940-3b96-41aa-8624-8338ee544fe1\") " pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.510948 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7brf9\" (UniqueName: \"kubernetes.io/projected/e8eb0c45-6539-478d-96ff-1d43d5d9978b-kube-api-access-7brf9\") pod \"machine-approver-56656f9798-p76w8\" (UID: \"e8eb0c45-6539-478d-96ff-1d43d5d9978b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p76w8" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.527597 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftgw2"] Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.528054 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-v676n"] Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.529167 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-kdwqh"] Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.534956 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ts95h\" (UniqueName: \"kubernetes.io/projected/2b720224-3e65-4443-a056-4e8fd226bcea-kube-api-access-ts95h\") pod \"apiserver-7bbb656c7d-bf4gt\" (UID: \"2b720224-3e65-4443-a056-4e8fd226bcea\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bf4gt" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.550811 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-wdjfz"] Sep 30 13:58:40 crc kubenswrapper[4840]: W0930 13:58:40.551391 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2842cd75_1ff5_4a3f_b7a0_8205388f4a17.slice/crio-e2521a8198db2bb35054a68c4f14f746a58a38ba30c028d7cd9c633400c0d9ef WatchSource:0}: Error finding container e2521a8198db2bb35054a68c4f14f746a58a38ba30c028d7cd9c633400c0d9ef: Status 404 returned error can't find the container with id e2521a8198db2bb35054a68c4f14f746a58a38ba30c028d7cd9c633400c0d9ef Sep 30 13:58:40 crc kubenswrapper[4840]: W0930 13:58:40.551694 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb0b27d1_5cb0_4a2c_95d7_012bc0a2d071.slice/crio-adc0252c3df50692250d0e0a4c0d5abb28b5b4add46d740876342a268ebaedab WatchSource:0}: Error finding container adc0252c3df50692250d0e0a4c0d5abb28b5b4add46d740876342a268ebaedab: Status 404 returned error can't find the container with id adc0252c3df50692250d0e0a4c0d5abb28b5b4add46d740876342a268ebaedab Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.557238 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-qjxnf"] Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.558574 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmbkh\" (UniqueName: \"kubernetes.io/projected/f0515176-8bf5-46a4-a905-7d308f64449e-kube-api-access-fmbkh\") pod \"router-default-5444994796-s9lx8\" (UID: \"f0515176-8bf5-46a4-a905-7d308f64449e\") " pod="openshift-ingress/router-default-5444994796-s9lx8" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.563238 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-dckvx" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.569711 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.572168 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0ba2ddf1-a9b7-448a-9e48-cdc9e80f3873-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-4942b\" (UID: \"0ba2ddf1-a9b7-448a-9e48-cdc9e80f3873\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4942b" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.577251 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-jfnms" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.591468 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8dpt\" (UniqueName: \"kubernetes.io/projected/ea0a1a12-942c-4ca3-b649-5ad87af478c8-kube-api-access-d8dpt\") pod \"openshift-controller-manager-operator-756b6f6bc6-sc7xq\" (UID: \"ea0a1a12-942c-4ca3-b649-5ad87af478c8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sc7xq" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.596849 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bf4gt" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.609455 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9rnrg"] Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.611178 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-s9lx8" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.629834 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gb9h8\" (UniqueName: \"kubernetes.io/projected/d0cf2421-0839-45da-bf4d-2c26e3137d16-kube-api-access-gb9h8\") pod \"openshift-apiserver-operator-796bbdcf4f-m2bq7\" (UID: \"d0cf2421-0839-45da-bf4d-2c26e3137d16\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m2bq7" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.648779 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdp5r\" (UniqueName: \"kubernetes.io/projected/8d0002f3-3315-4c27-bc39-2e5989b67d6b-kube-api-access-jdp5r\") pod \"openshift-config-operator-7777fb866f-9s79z\" (UID: \"8d0002f3-3315-4c27-bc39-2e5989b67d6b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9s79z" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.665313 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4942b" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.669680 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/46efdeea-a4d2-4ff5-898d-9ee516105cf4-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vckwz\" (UID: \"46efdeea-a4d2-4ff5-898d-9ee516105cf4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vckwz" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.670384 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-zbnls"] Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.671743 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmrwg\" (UniqueName: \"kubernetes.io/projected/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-kube-api-access-qmrwg\") pod \"oauth-openshift-558db77b4-ncblm\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:40 crc kubenswrapper[4840]: W0930 13:58:40.678733 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9dd57b46_77fc_4b8a_9fc1_a50084165914.slice/crio-d9412d6bc11d9a6ba329c315f49c51762fe98e4ea2c43e53b72fa624abc09c85 WatchSource:0}: Error finding container d9412d6bc11d9a6ba329c315f49c51762fe98e4ea2c43e53b72fa624abc09c85: Status 404 returned error can't find the container with id d9412d6bc11d9a6ba329c315f49c51762fe98e4ea2c43e53b72fa624abc09c85 Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.691321 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/11f26882-b509-4819-9fbb-2c69c062c951-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-lpb68\" (UID: \"11f26882-b509-4819-9fbb-2c69c062c951\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpb68" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.709609 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwj4n\" (UniqueName: \"kubernetes.io/projected/b4d78aa2-9ff6-404d-9b69-b8da8dfb2f9f-kube-api-access-vwj4n\") pod \"dns-operator-744455d44c-l47x7\" (UID: \"b4d78aa2-9ff6-404d-9b69-b8da8dfb2f9f\") " pod="openshift-dns-operator/dns-operator-744455d44c-l47x7" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.727858 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdvvd\" (UniqueName: \"kubernetes.io/projected/11f26882-b509-4819-9fbb-2c69c062c951-kube-api-access-qdvvd\") pod \"cluster-image-registry-operator-dc59b4c8b-lpb68\" (UID: \"11f26882-b509-4819-9fbb-2c69c062c951\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpb68" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.747810 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f71cd461-05df-4f5d-bf11-edf4d313996b-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-9szj8\" (UID: \"f71cd461-05df-4f5d-bf11-edf4d313996b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9szj8" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.751608 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p76w8" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.767455 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bx8vf\" (UniqueName: \"kubernetes.io/projected/94d7d8b4-fc87-48bc-ac61-84e39213d89a-kube-api-access-bx8vf\") pod \"dns-default-68gcw\" (UID: \"94d7d8b4-fc87-48bc-ac61-84e39213d89a\") " pod="openshift-dns/dns-default-68gcw" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.788182 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnq5m\" (UniqueName: \"kubernetes.io/projected/82fad711-fedd-452d-83a5-cf19fbf89815-kube-api-access-qnq5m\") pod \"machine-config-controller-84d6567774-8h224\" (UID: \"82fad711-fedd-452d-83a5-cf19fbf89815\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8h224" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.790890 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9s79z" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.801086 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.824478 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkk5g\" (UniqueName: \"kubernetes.io/projected/32d505eb-6035-488f-b154-2699e8af36ed-kube-api-access-qkk5g\") pod \"marketplace-operator-79b997595-snznw\" (UID: \"32d505eb-6035-488f-b154-2699e8af36ed\") " pod="openshift-marketplace/marketplace-operator-79b997595-snznw" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.832590 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-wdjfz" event={"ID":"661771aa-bd40-40fd-8a59-dc3b8156e5c5","Type":"ContainerStarted","Data":"6a56e0bfa14c27e4be3dc1dd85f6ea07a0b8d4189d81615588d76202ea3ef1a1"} Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.833591 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-zbnls" event={"ID":"d3508091-c5c6-4dca-9b32-883426154a6e","Type":"ContainerStarted","Data":"b2b5e548a14acf55a952ab07f95f1ecf5c07d21e8cc5075d689daac0be0e241a"} Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.834516 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-qjxnf" event={"ID":"9dd57b46-77fc-4b8a-9fc1-a50084165914","Type":"ContainerStarted","Data":"d9412d6bc11d9a6ba329c315f49c51762fe98e4ea2c43e53b72fa624abc09c85"} Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.835171 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26nwz\" (UniqueName: \"kubernetes.io/projected/69f4620c-5949-4527-aecb-9765f22e653e-kube-api-access-26nwz\") pod \"package-server-manager-789f6589d5-9sxdn\" (UID: \"69f4620c-5949-4527-aecb-9765f22e653e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9sxdn" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.835254 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-kdwqh" event={"ID":"dde0d17e-6bd0-4acb-b8b8-a8b0bc988a6e","Type":"ContainerStarted","Data":"aa1bd4f1a847b215126f494c068150c7039ca3354fc2d0780bc4ef8d0383d7e3"} Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.836813 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-v676n" event={"ID":"bb0b27d1-5cb0-4a2c-95d7-012bc0a2d071","Type":"ContainerStarted","Data":"adc0252c3df50692250d0e0a4c0d5abb28b5b4add46d740876342a268ebaedab"} Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.837943 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftgw2" event={"ID":"2842cd75-1ff5-4a3f-b7a0-8205388f4a17","Type":"ContainerStarted","Data":"e2521a8198db2bb35054a68c4f14f746a58a38ba30c028d7cd9c633400c0d9ef"} Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.846584 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m2bq7" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.855322 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-68gcw" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.856727 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rg7s\" (UniqueName: \"kubernetes.io/projected/713dae78-c3ff-40b4-a018-f801cf0562dd-kube-api-access-7rg7s\") pod \"multus-admission-controller-857f4d67dd-ptz4x\" (UID: \"713dae78-c3ff-40b4-a018-f801cf0562dd\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ptz4x" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.875671 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.883674 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sc7xq" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.890282 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpb68" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.896222 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.905673 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-l47x7" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.915899 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.917144 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vckwz" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.935666 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.973267 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f06465db-0dee-4f91-9e13-e46810c0e20e-signing-cabundle\") pod \"service-ca-9c57cc56f-6spsg\" (UID: \"f06465db-0dee-4f91-9e13-e46810c0e20e\") " pod="openshift-service-ca/service-ca-9c57cc56f-6spsg" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.973584 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-trusted-ca\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.973684 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7f88fdb6-dac3-4508-95a6-976cd084c4a7-srv-cert\") pod \"olm-operator-6b444d44fb-czbs2\" (UID: \"7f88fdb6-dac3-4508-95a6-976cd084c4a7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-czbs2" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.973834 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggmhx\" (UniqueName: \"kubernetes.io/projected/f06465db-0dee-4f91-9e13-e46810c0e20e-kube-api-access-ggmhx\") pod \"service-ca-9c57cc56f-6spsg\" (UID: \"f06465db-0dee-4f91-9e13-e46810c0e20e\") " pod="openshift-service-ca/service-ca-9c57cc56f-6spsg" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.973924 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2j2l4\" (UniqueName: \"kubernetes.io/projected/9955a9fc-27dc-4765-8bb9-e84fbb5448f1-kube-api-access-2j2l4\") pod \"kube-storage-version-migrator-operator-b67b599dd-kkmmf\" (UID: \"9955a9fc-27dc-4765-8bb9-e84fbb5448f1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kkmmf" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.974006 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a9361258-eda9-4ee7-8147-5a720baf1e86-srv-cert\") pod \"catalog-operator-68c6474976-s72m9\" (UID: \"a9361258-eda9-4ee7-8147-5a720baf1e86\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s72m9" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.974079 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-ca-trust-extracted\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.974152 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f06465db-0dee-4f91-9e13-e46810c0e20e-signing-key\") pod \"service-ca-9c57cc56f-6spsg\" (UID: \"f06465db-0dee-4f91-9e13-e46810c0e20e\") " pod="openshift-service-ca/service-ca-9c57cc56f-6spsg" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.974239 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.974337 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km26r\" (UniqueName: \"kubernetes.io/projected/771e14ef-523d-4eab-b13b-87a112fa11bf-kube-api-access-km26r\") pod \"machine-config-operator-74547568cd-dksqg\" (UID: \"771e14ef-523d-4eab-b13b-87a112fa11bf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dksqg" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.974453 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5n76\" (UniqueName: \"kubernetes.io/projected/7f88fdb6-dac3-4508-95a6-976cd084c4a7-kube-api-access-b5n76\") pod \"olm-operator-6b444d44fb-czbs2\" (UID: \"7f88fdb6-dac3-4508-95a6-976cd084c4a7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-czbs2" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.974524 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-registry-tls\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.974619 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7f88fdb6-dac3-4508-95a6-976cd084c4a7-profile-collector-cert\") pod \"olm-operator-6b444d44fb-czbs2\" (UID: \"7f88fdb6-dac3-4508-95a6-976cd084c4a7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-czbs2" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.974698 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/771e14ef-523d-4eab-b13b-87a112fa11bf-proxy-tls\") pod \"machine-config-operator-74547568cd-dksqg\" (UID: \"771e14ef-523d-4eab-b13b-87a112fa11bf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dksqg" Sep 30 13:58:40 crc kubenswrapper[4840]: E0930 13:58:40.974774 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:41.474762644 +0000 UTC m=+150.103849067 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.974811 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-bound-sa-token\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.974831 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgclp\" (UniqueName: \"kubernetes.io/projected/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-kube-api-access-fgclp\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.974868 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkrg5\" (UniqueName: \"kubernetes.io/projected/850a2c18-882c-495f-a7e8-743e425a8b61-kube-api-access-gkrg5\") pod \"collect-profiles-29320665-rkbh5\" (UID: \"850a2c18-882c-495f-a7e8-743e425a8b61\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320665-rkbh5" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.975184 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8fe4d47a-5be0-4130-92e7-da8b26b8940b-trusted-ca\") pod \"ingress-operator-5b745b69d9-bgjpl\" (UID: \"8fe4d47a-5be0-4130-92e7-da8b26b8940b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bgjpl" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.975303 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/eb59a261-a39c-4be5-86f0-a14643aef960-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-z866q\" (UID: \"eb59a261-a39c-4be5-86f0-a14643aef960\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z866q" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.975414 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9955a9fc-27dc-4765-8bb9-e84fbb5448f1-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-kkmmf\" (UID: \"9955a9fc-27dc-4765-8bb9-e84fbb5448f1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kkmmf" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.975624 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7c43971-89ed-4b19-8138-0baced4bdd1f-config\") pod \"service-ca-operator-777779d784-w4vtf\" (UID: \"f7c43971-89ed-4b19-8138-0baced4bdd1f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-w4vtf" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.975850 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/850a2c18-882c-495f-a7e8-743e425a8b61-config-volume\") pod \"collect-profiles-29320665-rkbh5\" (UID: \"850a2c18-882c-495f-a7e8-743e425a8b61\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320665-rkbh5" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.975957 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkblt\" (UniqueName: \"kubernetes.io/projected/eb59a261-a39c-4be5-86f0-a14643aef960-kube-api-access-hkblt\") pod \"control-plane-machine-set-operator-78cbb6b69f-z866q\" (UID: \"eb59a261-a39c-4be5-86f0-a14643aef960\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z866q" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.976097 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f5ac3e36-a48a-477a-b83e-6d9acaba6169-apiservice-cert\") pod \"packageserver-d55dfcdfc-n2cvx\" (UID: \"f5ac3e36-a48a-477a-b83e-6d9acaba6169\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n2cvx" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.976160 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5dmt\" (UniqueName: \"kubernetes.io/projected/f5ac3e36-a48a-477a-b83e-6d9acaba6169-kube-api-access-x5dmt\") pod \"packageserver-d55dfcdfc-n2cvx\" (UID: \"f5ac3e36-a48a-477a-b83e-6d9acaba6169\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n2cvx" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.976183 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-installation-pull-secrets\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.976210 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7c43971-89ed-4b19-8138-0baced4bdd1f-serving-cert\") pod \"service-ca-operator-777779d784-w4vtf\" (UID: \"f7c43971-89ed-4b19-8138-0baced4bdd1f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-w4vtf" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.976248 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9szj8" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.976387 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a9361258-eda9-4ee7-8147-5a720baf1e86-profile-collector-cert\") pod \"catalog-operator-68c6474976-s72m9\" (UID: \"a9361258-eda9-4ee7-8147-5a720baf1e86\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s72m9" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.976597 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8fe4d47a-5be0-4130-92e7-da8b26b8940b-bound-sa-token\") pod \"ingress-operator-5b745b69d9-bgjpl\" (UID: \"8fe4d47a-5be0-4130-92e7-da8b26b8940b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bgjpl" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.977435 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/771e14ef-523d-4eab-b13b-87a112fa11bf-auth-proxy-config\") pod \"machine-config-operator-74547568cd-dksqg\" (UID: \"771e14ef-523d-4eab-b13b-87a112fa11bf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dksqg" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.977465 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mb5fg\" (UniqueName: \"kubernetes.io/projected/1454044a-1464-42f3-92c4-f6093e3bd7d1-kube-api-access-mb5fg\") pod \"migrator-59844c95c7-jslz5\" (UID: \"1454044a-1464-42f3-92c4-f6093e3bd7d1\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-jslz5" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.977497 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/f5ac3e36-a48a-477a-b83e-6d9acaba6169-tmpfs\") pod \"packageserver-d55dfcdfc-n2cvx\" (UID: \"f5ac3e36-a48a-477a-b83e-6d9acaba6169\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n2cvx" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.977614 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-registry-certificates\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.977696 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/850a2c18-882c-495f-a7e8-743e425a8b61-secret-volume\") pod \"collect-profiles-29320665-rkbh5\" (UID: \"850a2c18-882c-495f-a7e8-743e425a8b61\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320665-rkbh5" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.977726 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f5ac3e36-a48a-477a-b83e-6d9acaba6169-webhook-cert\") pod \"packageserver-d55dfcdfc-n2cvx\" (UID: \"f5ac3e36-a48a-477a-b83e-6d9acaba6169\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n2cvx" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.977754 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2wfp\" (UniqueName: \"kubernetes.io/projected/a9361258-eda9-4ee7-8147-5a720baf1e86-kube-api-access-z2wfp\") pod \"catalog-operator-68c6474976-s72m9\" (UID: \"a9361258-eda9-4ee7-8147-5a720baf1e86\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s72m9" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.977814 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8fe4d47a-5be0-4130-92e7-da8b26b8940b-metrics-tls\") pod \"ingress-operator-5b745b69d9-bgjpl\" (UID: \"8fe4d47a-5be0-4130-92e7-da8b26b8940b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bgjpl" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.977848 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9955a9fc-27dc-4765-8bb9-e84fbb5448f1-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-kkmmf\" (UID: \"9955a9fc-27dc-4765-8bb9-e84fbb5448f1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kkmmf" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.977876 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/771e14ef-523d-4eab-b13b-87a112fa11bf-images\") pod \"machine-config-operator-74547568cd-dksqg\" (UID: \"771e14ef-523d-4eab-b13b-87a112fa11bf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dksqg" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.977904 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpnkx\" (UniqueName: \"kubernetes.io/projected/8fe4d47a-5be0-4130-92e7-da8b26b8940b-kube-api-access-cpnkx\") pod \"ingress-operator-5b745b69d9-bgjpl\" (UID: \"8fe4d47a-5be0-4130-92e7-da8b26b8940b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bgjpl" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.977964 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jx98f\" (UniqueName: \"kubernetes.io/projected/f7c43971-89ed-4b19-8138-0baced4bdd1f-kube-api-access-jx98f\") pod \"service-ca-operator-777779d784-w4vtf\" (UID: \"f7c43971-89ed-4b19-8138-0baced4bdd1f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-w4vtf" Sep 30 13:58:40 crc kubenswrapper[4840]: I0930 13:58:40.997719 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-ptz4x" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.038169 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-snznw" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.071780 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8h224" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.079474 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:41 crc kubenswrapper[4840]: E0930 13:58:41.079663 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:41.579614405 +0000 UTC m=+150.208700848 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.079704 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-trusted-ca\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.079776 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/c4ecdaa2-b6f9-4682-b115-eb12a0060263-csi-data-dir\") pod \"csi-hostpathplugin-56p76\" (UID: \"c4ecdaa2-b6f9-4682-b115-eb12a0060263\") " pod="hostpath-provisioner/csi-hostpathplugin-56p76" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.079857 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7f88fdb6-dac3-4508-95a6-976cd084c4a7-srv-cert\") pod \"olm-operator-6b444d44fb-czbs2\" (UID: \"7f88fdb6-dac3-4508-95a6-976cd084c4a7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-czbs2" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.079929 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggmhx\" (UniqueName: \"kubernetes.io/projected/f06465db-0dee-4f91-9e13-e46810c0e20e-kube-api-access-ggmhx\") pod \"service-ca-9c57cc56f-6spsg\" (UID: \"f06465db-0dee-4f91-9e13-e46810c0e20e\") " pod="openshift-service-ca/service-ca-9c57cc56f-6spsg" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.079967 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2j2l4\" (UniqueName: \"kubernetes.io/projected/9955a9fc-27dc-4765-8bb9-e84fbb5448f1-kube-api-access-2j2l4\") pod \"kube-storage-version-migrator-operator-b67b599dd-kkmmf\" (UID: \"9955a9fc-27dc-4765-8bb9-e84fbb5448f1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kkmmf" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.080041 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a9361258-eda9-4ee7-8147-5a720baf1e86-srv-cert\") pod \"catalog-operator-68c6474976-s72m9\" (UID: \"a9361258-eda9-4ee7-8147-5a720baf1e86\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s72m9" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.080099 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/c4ecdaa2-b6f9-4682-b115-eb12a0060263-mountpoint-dir\") pod \"csi-hostpathplugin-56p76\" (UID: \"c4ecdaa2-b6f9-4682-b115-eb12a0060263\") " pod="hostpath-provisioner/csi-hostpathplugin-56p76" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.080129 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfs2k\" (UniqueName: \"kubernetes.io/projected/c4ecdaa2-b6f9-4682-b115-eb12a0060263-kube-api-access-lfs2k\") pod \"csi-hostpathplugin-56p76\" (UID: \"c4ecdaa2-b6f9-4682-b115-eb12a0060263\") " pod="hostpath-provisioner/csi-hostpathplugin-56p76" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.080200 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b526l\" (UniqueName: \"kubernetes.io/projected/3869773f-0c18-4f5d-a0e8-2f35ad030b5c-kube-api-access-b526l\") pod \"ingress-canary-8j25v\" (UID: \"3869773f-0c18-4f5d-a0e8-2f35ad030b5c\") " pod="openshift-ingress-canary/ingress-canary-8j25v" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.080297 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.080369 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-ca-trust-extracted\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.080401 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f06465db-0dee-4f91-9e13-e46810c0e20e-signing-key\") pod \"service-ca-9c57cc56f-6spsg\" (UID: \"f06465db-0dee-4f91-9e13-e46810c0e20e\") " pod="openshift-service-ca/service-ca-9c57cc56f-6spsg" Sep 30 13:58:41 crc kubenswrapper[4840]: E0930 13:58:41.080961 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:41.58093456 +0000 UTC m=+150.210021023 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.081144 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km26r\" (UniqueName: \"kubernetes.io/projected/771e14ef-523d-4eab-b13b-87a112fa11bf-kube-api-access-km26r\") pod \"machine-config-operator-74547568cd-dksqg\" (UID: \"771e14ef-523d-4eab-b13b-87a112fa11bf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dksqg" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.081265 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5n76\" (UniqueName: \"kubernetes.io/projected/7f88fdb6-dac3-4508-95a6-976cd084c4a7-kube-api-access-b5n76\") pod \"olm-operator-6b444d44fb-czbs2\" (UID: \"7f88fdb6-dac3-4508-95a6-976cd084c4a7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-czbs2" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.081305 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/771e14ef-523d-4eab-b13b-87a112fa11bf-proxy-tls\") pod \"machine-config-operator-74547568cd-dksqg\" (UID: \"771e14ef-523d-4eab-b13b-87a112fa11bf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dksqg" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.081360 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-registry-tls\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.081382 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7f88fdb6-dac3-4508-95a6-976cd084c4a7-profile-collector-cert\") pod \"olm-operator-6b444d44fb-czbs2\" (UID: \"7f88fdb6-dac3-4508-95a6-976cd084c4a7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-czbs2" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.081414 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3869773f-0c18-4f5d-a0e8-2f35ad030b5c-cert\") pod \"ingress-canary-8j25v\" (UID: \"3869773f-0c18-4f5d-a0e8-2f35ad030b5c\") " pod="openshift-ingress-canary/ingress-canary-8j25v" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.081477 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-bound-sa-token\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.081502 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgclp\" (UniqueName: \"kubernetes.io/projected/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-kube-api-access-fgclp\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.081569 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkrg5\" (UniqueName: \"kubernetes.io/projected/850a2c18-882c-495f-a7e8-743e425a8b61-kube-api-access-gkrg5\") pod \"collect-profiles-29320665-rkbh5\" (UID: \"850a2c18-882c-495f-a7e8-743e425a8b61\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320665-rkbh5" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.081810 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-ca-trust-extracted\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.081626 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8fe4d47a-5be0-4130-92e7-da8b26b8940b-trusted-ca\") pod \"ingress-operator-5b745b69d9-bgjpl\" (UID: \"8fe4d47a-5be0-4130-92e7-da8b26b8940b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bgjpl" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.082321 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/eb59a261-a39c-4be5-86f0-a14643aef960-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-z866q\" (UID: \"eb59a261-a39c-4be5-86f0-a14643aef960\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z866q" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.082353 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9955a9fc-27dc-4765-8bb9-e84fbb5448f1-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-kkmmf\" (UID: \"9955a9fc-27dc-4765-8bb9-e84fbb5448f1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kkmmf" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.082386 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/c4ecdaa2-b6f9-4682-b115-eb12a0060263-registration-dir\") pod \"csi-hostpathplugin-56p76\" (UID: \"c4ecdaa2-b6f9-4682-b115-eb12a0060263\") " pod="hostpath-provisioner/csi-hostpathplugin-56p76" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.082412 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7c43971-89ed-4b19-8138-0baced4bdd1f-config\") pod \"service-ca-operator-777779d784-w4vtf\" (UID: \"f7c43971-89ed-4b19-8138-0baced4bdd1f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-w4vtf" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.082454 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/c4ecdaa2-b6f9-4682-b115-eb12a0060263-plugins-dir\") pod \"csi-hostpathplugin-56p76\" (UID: \"c4ecdaa2-b6f9-4682-b115-eb12a0060263\") " pod="hostpath-provisioner/csi-hostpathplugin-56p76" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.082484 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/850a2c18-882c-495f-a7e8-743e425a8b61-config-volume\") pod \"collect-profiles-29320665-rkbh5\" (UID: \"850a2c18-882c-495f-a7e8-743e425a8b61\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320665-rkbh5" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.082517 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkblt\" (UniqueName: \"kubernetes.io/projected/eb59a261-a39c-4be5-86f0-a14643aef960-kube-api-access-hkblt\") pod \"control-plane-machine-set-operator-78cbb6b69f-z866q\" (UID: \"eb59a261-a39c-4be5-86f0-a14643aef960\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z866q" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.082582 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f5ac3e36-a48a-477a-b83e-6d9acaba6169-apiservice-cert\") pod \"packageserver-d55dfcdfc-n2cvx\" (UID: \"f5ac3e36-a48a-477a-b83e-6d9acaba6169\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n2cvx" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.082626 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-trusted-ca\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.082762 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5dmt\" (UniqueName: \"kubernetes.io/projected/f5ac3e36-a48a-477a-b83e-6d9acaba6169-kube-api-access-x5dmt\") pod \"packageserver-d55dfcdfc-n2cvx\" (UID: \"f5ac3e36-a48a-477a-b83e-6d9acaba6169\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n2cvx" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.082800 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-installation-pull-secrets\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.082866 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7c43971-89ed-4b19-8138-0baced4bdd1f-serving-cert\") pod \"service-ca-operator-777779d784-w4vtf\" (UID: \"f7c43971-89ed-4b19-8138-0baced4bdd1f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-w4vtf" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.083070 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/641b86bf-c833-495c-9444-0c3c28bac9cc-certs\") pod \"machine-config-server-zb5zq\" (UID: \"641b86bf-c833-495c-9444-0c3c28bac9cc\") " pod="openshift-machine-config-operator/machine-config-server-zb5zq" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.083195 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8fe4d47a-5be0-4130-92e7-da8b26b8940b-bound-sa-token\") pod \"ingress-operator-5b745b69d9-bgjpl\" (UID: \"8fe4d47a-5be0-4130-92e7-da8b26b8940b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bgjpl" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.083230 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a9361258-eda9-4ee7-8147-5a720baf1e86-profile-collector-cert\") pod \"catalog-operator-68c6474976-s72m9\" (UID: \"a9361258-eda9-4ee7-8147-5a720baf1e86\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s72m9" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.083267 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/771e14ef-523d-4eab-b13b-87a112fa11bf-auth-proxy-config\") pod \"machine-config-operator-74547568cd-dksqg\" (UID: \"771e14ef-523d-4eab-b13b-87a112fa11bf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dksqg" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.083299 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mb5fg\" (UniqueName: \"kubernetes.io/projected/1454044a-1464-42f3-92c4-f6093e3bd7d1-kube-api-access-mb5fg\") pod \"migrator-59844c95c7-jslz5\" (UID: \"1454044a-1464-42f3-92c4-f6093e3bd7d1\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-jslz5" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.084033 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8fe4d47a-5be0-4130-92e7-da8b26b8940b-trusted-ca\") pod \"ingress-operator-5b745b69d9-bgjpl\" (UID: \"8fe4d47a-5be0-4130-92e7-da8b26b8940b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bgjpl" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.084110 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xxjf\" (UniqueName: \"kubernetes.io/projected/641b86bf-c833-495c-9444-0c3c28bac9cc-kube-api-access-4xxjf\") pod \"machine-config-server-zb5zq\" (UID: \"641b86bf-c833-495c-9444-0c3c28bac9cc\") " pod="openshift-machine-config-operator/machine-config-server-zb5zq" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.084139 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9955a9fc-27dc-4765-8bb9-e84fbb5448f1-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-kkmmf\" (UID: \"9955a9fc-27dc-4765-8bb9-e84fbb5448f1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kkmmf" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.084263 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/f5ac3e36-a48a-477a-b83e-6d9acaba6169-tmpfs\") pod \"packageserver-d55dfcdfc-n2cvx\" (UID: \"f5ac3e36-a48a-477a-b83e-6d9acaba6169\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n2cvx" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.084711 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-registry-certificates\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.085452 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/850a2c18-882c-495f-a7e8-743e425a8b61-config-volume\") pod \"collect-profiles-29320665-rkbh5\" (UID: \"850a2c18-882c-495f-a7e8-743e425a8b61\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320665-rkbh5" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.085682 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7f88fdb6-dac3-4508-95a6-976cd084c4a7-srv-cert\") pod \"olm-operator-6b444d44fb-czbs2\" (UID: \"7f88fdb6-dac3-4508-95a6-976cd084c4a7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-czbs2" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.085883 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/850a2c18-882c-495f-a7e8-743e425a8b61-secret-volume\") pod \"collect-profiles-29320665-rkbh5\" (UID: \"850a2c18-882c-495f-a7e8-743e425a8b61\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320665-rkbh5" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.086091 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/c4ecdaa2-b6f9-4682-b115-eb12a0060263-socket-dir\") pod \"csi-hostpathplugin-56p76\" (UID: \"c4ecdaa2-b6f9-4682-b115-eb12a0060263\") " pod="hostpath-provisioner/csi-hostpathplugin-56p76" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.086138 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8fe4d47a-5be0-4130-92e7-da8b26b8940b-metrics-tls\") pod \"ingress-operator-5b745b69d9-bgjpl\" (UID: \"8fe4d47a-5be0-4130-92e7-da8b26b8940b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bgjpl" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.086201 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9955a9fc-27dc-4765-8bb9-e84fbb5448f1-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-kkmmf\" (UID: \"9955a9fc-27dc-4765-8bb9-e84fbb5448f1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kkmmf" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.086245 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f5ac3e36-a48a-477a-b83e-6d9acaba6169-webhook-cert\") pod \"packageserver-d55dfcdfc-n2cvx\" (UID: \"f5ac3e36-a48a-477a-b83e-6d9acaba6169\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n2cvx" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.086271 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2wfp\" (UniqueName: \"kubernetes.io/projected/a9361258-eda9-4ee7-8147-5a720baf1e86-kube-api-access-z2wfp\") pod \"catalog-operator-68c6474976-s72m9\" (UID: \"a9361258-eda9-4ee7-8147-5a720baf1e86\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s72m9" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.086294 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/771e14ef-523d-4eab-b13b-87a112fa11bf-images\") pod \"machine-config-operator-74547568cd-dksqg\" (UID: \"771e14ef-523d-4eab-b13b-87a112fa11bf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dksqg" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.086324 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpnkx\" (UniqueName: \"kubernetes.io/projected/8fe4d47a-5be0-4130-92e7-da8b26b8940b-kube-api-access-cpnkx\") pod \"ingress-operator-5b745b69d9-bgjpl\" (UID: \"8fe4d47a-5be0-4130-92e7-da8b26b8940b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bgjpl" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.086370 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/641b86bf-c833-495c-9444-0c3c28bac9cc-node-bootstrap-token\") pod \"machine-config-server-zb5zq\" (UID: \"641b86bf-c833-495c-9444-0c3c28bac9cc\") " pod="openshift-machine-config-operator/machine-config-server-zb5zq" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.086400 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jx98f\" (UniqueName: \"kubernetes.io/projected/f7c43971-89ed-4b19-8138-0baced4bdd1f-kube-api-access-jx98f\") pod \"service-ca-operator-777779d784-w4vtf\" (UID: \"f7c43971-89ed-4b19-8138-0baced4bdd1f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-w4vtf" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.086444 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f06465db-0dee-4f91-9e13-e46810c0e20e-signing-cabundle\") pod \"service-ca-9c57cc56f-6spsg\" (UID: \"f06465db-0dee-4f91-9e13-e46810c0e20e\") " pod="openshift-service-ca/service-ca-9c57cc56f-6spsg" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.086580 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/f5ac3e36-a48a-477a-b83e-6d9acaba6169-tmpfs\") pod \"packageserver-d55dfcdfc-n2cvx\" (UID: \"f5ac3e36-a48a-477a-b83e-6d9acaba6169\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n2cvx" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.087580 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-registry-tls\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.088059 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f06465db-0dee-4f91-9e13-e46810c0e20e-signing-key\") pod \"service-ca-9c57cc56f-6spsg\" (UID: \"f06465db-0dee-4f91-9e13-e46810c0e20e\") " pod="openshift-service-ca/service-ca-9c57cc56f-6spsg" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.088518 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/771e14ef-523d-4eab-b13b-87a112fa11bf-images\") pod \"machine-config-operator-74547568cd-dksqg\" (UID: \"771e14ef-523d-4eab-b13b-87a112fa11bf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dksqg" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.088633 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/771e14ef-523d-4eab-b13b-87a112fa11bf-auth-proxy-config\") pod \"machine-config-operator-74547568cd-dksqg\" (UID: \"771e14ef-523d-4eab-b13b-87a112fa11bf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dksqg" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.088982 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a9361258-eda9-4ee7-8147-5a720baf1e86-srv-cert\") pod \"catalog-operator-68c6474976-s72m9\" (UID: \"a9361258-eda9-4ee7-8147-5a720baf1e86\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s72m9" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.089309 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7c43971-89ed-4b19-8138-0baced4bdd1f-config\") pod \"service-ca-operator-777779d784-w4vtf\" (UID: \"f7c43971-89ed-4b19-8138-0baced4bdd1f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-w4vtf" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.089530 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/eb59a261-a39c-4be5-86f0-a14643aef960-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-z866q\" (UID: \"eb59a261-a39c-4be5-86f0-a14643aef960\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z866q" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.089702 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f5ac3e36-a48a-477a-b83e-6d9acaba6169-apiservice-cert\") pod \"packageserver-d55dfcdfc-n2cvx\" (UID: \"f5ac3e36-a48a-477a-b83e-6d9acaba6169\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n2cvx" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.089807 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7f88fdb6-dac3-4508-95a6-976cd084c4a7-profile-collector-cert\") pod \"olm-operator-6b444d44fb-czbs2\" (UID: \"7f88fdb6-dac3-4508-95a6-976cd084c4a7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-czbs2" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.090062 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-registry-certificates\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.090900 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f5ac3e36-a48a-477a-b83e-6d9acaba6169-webhook-cert\") pod \"packageserver-d55dfcdfc-n2cvx\" (UID: \"f5ac3e36-a48a-477a-b83e-6d9acaba6169\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n2cvx" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.091076 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-installation-pull-secrets\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.091351 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/850a2c18-882c-495f-a7e8-743e425a8b61-secret-volume\") pod \"collect-profiles-29320665-rkbh5\" (UID: \"850a2c18-882c-495f-a7e8-743e425a8b61\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320665-rkbh5" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.091868 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/771e14ef-523d-4eab-b13b-87a112fa11bf-proxy-tls\") pod \"machine-config-operator-74547568cd-dksqg\" (UID: \"771e14ef-523d-4eab-b13b-87a112fa11bf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dksqg" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.091924 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9955a9fc-27dc-4765-8bb9-e84fbb5448f1-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-kkmmf\" (UID: \"9955a9fc-27dc-4765-8bb9-e84fbb5448f1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kkmmf" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.093574 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a9361258-eda9-4ee7-8147-5a720baf1e86-profile-collector-cert\") pod \"catalog-operator-68c6474976-s72m9\" (UID: \"a9361258-eda9-4ee7-8147-5a720baf1e86\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s72m9" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.094759 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9sxdn" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.094868 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7c43971-89ed-4b19-8138-0baced4bdd1f-serving-cert\") pod \"service-ca-operator-777779d784-w4vtf\" (UID: \"f7c43971-89ed-4b19-8138-0baced4bdd1f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-w4vtf" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.132460 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2j2l4\" (UniqueName: \"kubernetes.io/projected/9955a9fc-27dc-4765-8bb9-e84fbb5448f1-kube-api-access-2j2l4\") pod \"kube-storage-version-migrator-operator-b67b599dd-kkmmf\" (UID: \"9955a9fc-27dc-4765-8bb9-e84fbb5448f1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kkmmf" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.148372 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggmhx\" (UniqueName: \"kubernetes.io/projected/f06465db-0dee-4f91-9e13-e46810c0e20e-kube-api-access-ggmhx\") pod \"service-ca-9c57cc56f-6spsg\" (UID: \"f06465db-0dee-4f91-9e13-e46810c0e20e\") " pod="openshift-service-ca/service-ca-9c57cc56f-6spsg" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.165752 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f06465db-0dee-4f91-9e13-e46810c0e20e-signing-cabundle\") pod \"service-ca-9c57cc56f-6spsg\" (UID: \"f06465db-0dee-4f91-9e13-e46810c0e20e\") " pod="openshift-service-ca/service-ca-9c57cc56f-6spsg" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.166631 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8fe4d47a-5be0-4130-92e7-da8b26b8940b-metrics-tls\") pod \"ingress-operator-5b745b69d9-bgjpl\" (UID: \"8fe4d47a-5be0-4130-92e7-da8b26b8940b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bgjpl" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.170193 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km26r\" (UniqueName: \"kubernetes.io/projected/771e14ef-523d-4eab-b13b-87a112fa11bf-kube-api-access-km26r\") pod \"machine-config-operator-74547568cd-dksqg\" (UID: \"771e14ef-523d-4eab-b13b-87a112fa11bf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dksqg" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.187118 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:41 crc kubenswrapper[4840]: E0930 13:58:41.187287 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:41.68726158 +0000 UTC m=+150.316348003 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.187383 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3869773f-0c18-4f5d-a0e8-2f35ad030b5c-cert\") pod \"ingress-canary-8j25v\" (UID: \"3869773f-0c18-4f5d-a0e8-2f35ad030b5c\") " pod="openshift-ingress-canary/ingress-canary-8j25v" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.187477 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/c4ecdaa2-b6f9-4682-b115-eb12a0060263-registration-dir\") pod \"csi-hostpathplugin-56p76\" (UID: \"c4ecdaa2-b6f9-4682-b115-eb12a0060263\") " pod="hostpath-provisioner/csi-hostpathplugin-56p76" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.187513 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/c4ecdaa2-b6f9-4682-b115-eb12a0060263-plugins-dir\") pod \"csi-hostpathplugin-56p76\" (UID: \"c4ecdaa2-b6f9-4682-b115-eb12a0060263\") " pod="hostpath-provisioner/csi-hostpathplugin-56p76" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.187610 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/641b86bf-c833-495c-9444-0c3c28bac9cc-certs\") pod \"machine-config-server-zb5zq\" (UID: \"641b86bf-c833-495c-9444-0c3c28bac9cc\") " pod="openshift-machine-config-operator/machine-config-server-zb5zq" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.187656 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xxjf\" (UniqueName: \"kubernetes.io/projected/641b86bf-c833-495c-9444-0c3c28bac9cc-kube-api-access-4xxjf\") pod \"machine-config-server-zb5zq\" (UID: \"641b86bf-c833-495c-9444-0c3c28bac9cc\") " pod="openshift-machine-config-operator/machine-config-server-zb5zq" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.187697 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/c4ecdaa2-b6f9-4682-b115-eb12a0060263-socket-dir\") pod \"csi-hostpathplugin-56p76\" (UID: \"c4ecdaa2-b6f9-4682-b115-eb12a0060263\") " pod="hostpath-provisioner/csi-hostpathplugin-56p76" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.187743 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/641b86bf-c833-495c-9444-0c3c28bac9cc-node-bootstrap-token\") pod \"machine-config-server-zb5zq\" (UID: \"641b86bf-c833-495c-9444-0c3c28bac9cc\") " pod="openshift-machine-config-operator/machine-config-server-zb5zq" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.187794 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/c4ecdaa2-b6f9-4682-b115-eb12a0060263-csi-data-dir\") pod \"csi-hostpathplugin-56p76\" (UID: \"c4ecdaa2-b6f9-4682-b115-eb12a0060263\") " pod="hostpath-provisioner/csi-hostpathplugin-56p76" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.187831 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/c4ecdaa2-b6f9-4682-b115-eb12a0060263-mountpoint-dir\") pod \"csi-hostpathplugin-56p76\" (UID: \"c4ecdaa2-b6f9-4682-b115-eb12a0060263\") " pod="hostpath-provisioner/csi-hostpathplugin-56p76" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.187854 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfs2k\" (UniqueName: \"kubernetes.io/projected/c4ecdaa2-b6f9-4682-b115-eb12a0060263-kube-api-access-lfs2k\") pod \"csi-hostpathplugin-56p76\" (UID: \"c4ecdaa2-b6f9-4682-b115-eb12a0060263\") " pod="hostpath-provisioner/csi-hostpathplugin-56p76" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.187867 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/c4ecdaa2-b6f9-4682-b115-eb12a0060263-plugins-dir\") pod \"csi-hostpathplugin-56p76\" (UID: \"c4ecdaa2-b6f9-4682-b115-eb12a0060263\") " pod="hostpath-provisioner/csi-hostpathplugin-56p76" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.187884 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.187938 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b526l\" (UniqueName: \"kubernetes.io/projected/3869773f-0c18-4f5d-a0e8-2f35ad030b5c-kube-api-access-b526l\") pod \"ingress-canary-8j25v\" (UID: \"3869773f-0c18-4f5d-a0e8-2f35ad030b5c\") " pod="openshift-ingress-canary/ingress-canary-8j25v" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.187868 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/c4ecdaa2-b6f9-4682-b115-eb12a0060263-socket-dir\") pod \"csi-hostpathplugin-56p76\" (UID: \"c4ecdaa2-b6f9-4682-b115-eb12a0060263\") " pod="hostpath-provisioner/csi-hostpathplugin-56p76" Sep 30 13:58:41 crc kubenswrapper[4840]: E0930 13:58:41.188173 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:41.688157804 +0000 UTC m=+150.317244317 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.188291 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/c4ecdaa2-b6f9-4682-b115-eb12a0060263-mountpoint-dir\") pod \"csi-hostpathplugin-56p76\" (UID: \"c4ecdaa2-b6f9-4682-b115-eb12a0060263\") " pod="hostpath-provisioner/csi-hostpathplugin-56p76" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.188477 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/c4ecdaa2-b6f9-4682-b115-eb12a0060263-registration-dir\") pod \"csi-hostpathplugin-56p76\" (UID: \"c4ecdaa2-b6f9-4682-b115-eb12a0060263\") " pod="hostpath-provisioner/csi-hostpathplugin-56p76" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.188883 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/c4ecdaa2-b6f9-4682-b115-eb12a0060263-csi-data-dir\") pod \"csi-hostpathplugin-56p76\" (UID: \"c4ecdaa2-b6f9-4682-b115-eb12a0060263\") " pod="hostpath-provisioner/csi-hostpathplugin-56p76" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.191575 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/641b86bf-c833-495c-9444-0c3c28bac9cc-certs\") pod \"machine-config-server-zb5zq\" (UID: \"641b86bf-c833-495c-9444-0c3c28bac9cc\") " pod="openshift-machine-config-operator/machine-config-server-zb5zq" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.192285 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3869773f-0c18-4f5d-a0e8-2f35ad030b5c-cert\") pod \"ingress-canary-8j25v\" (UID: \"3869773f-0c18-4f5d-a0e8-2f35ad030b5c\") " pod="openshift-ingress-canary/ingress-canary-8j25v" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.192462 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgclp\" (UniqueName: \"kubernetes.io/projected/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-kube-api-access-fgclp\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.197694 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/641b86bf-c833-495c-9444-0c3c28bac9cc-node-bootstrap-token\") pod \"machine-config-server-zb5zq\" (UID: \"641b86bf-c833-495c-9444-0c3c28bac9cc\") " pod="openshift-machine-config-operator/machine-config-server-zb5zq" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.218400 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-bound-sa-token\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.231515 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5dmt\" (UniqueName: \"kubernetes.io/projected/f5ac3e36-a48a-477a-b83e-6d9acaba6169-kube-api-access-x5dmt\") pod \"packageserver-d55dfcdfc-n2cvx\" (UID: \"f5ac3e36-a48a-477a-b83e-6d9acaba6169\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n2cvx" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.245274 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-dckvx"] Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.250599 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkrg5\" (UniqueName: \"kubernetes.io/projected/850a2c18-882c-495f-a7e8-743e425a8b61-kube-api-access-gkrg5\") pod \"collect-profiles-29320665-rkbh5\" (UID: \"850a2c18-882c-495f-a7e8-743e425a8b61\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320665-rkbh5" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.270345 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mb5fg\" (UniqueName: \"kubernetes.io/projected/1454044a-1464-42f3-92c4-f6093e3bd7d1-kube-api-access-mb5fg\") pod \"migrator-59844c95c7-jslz5\" (UID: \"1454044a-1464-42f3-92c4-f6093e3bd7d1\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-jslz5" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.289115 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:41 crc kubenswrapper[4840]: E0930 13:58:41.289272 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:41.789250335 +0000 UTC m=+150.418336768 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.289465 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8fe4d47a-5be0-4130-92e7-da8b26b8940b-bound-sa-token\") pod \"ingress-operator-5b745b69d9-bgjpl\" (UID: \"8fe4d47a-5be0-4130-92e7-da8b26b8940b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bgjpl" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.289728 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:41 crc kubenswrapper[4840]: E0930 13:58:41.290056 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:41.790046166 +0000 UTC m=+150.419132589 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.303836 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kkmmf" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.309085 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkblt\" (UniqueName: \"kubernetes.io/projected/eb59a261-a39c-4be5-86f0-a14643aef960-kube-api-access-hkblt\") pod \"control-plane-machine-set-operator-78cbb6b69f-z866q\" (UID: \"eb59a261-a39c-4be5-86f0-a14643aef960\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z866q" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.311329 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-9s79z"] Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.311768 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-jslz5" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.318958 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n2cvx" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.328425 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5n76\" (UniqueName: \"kubernetes.io/projected/7f88fdb6-dac3-4508-95a6-976cd084c4a7-kube-api-access-b5n76\") pod \"olm-operator-6b444d44fb-czbs2\" (UID: \"7f88fdb6-dac3-4508-95a6-976cd084c4a7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-czbs2" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.331085 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dksqg" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.345008 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z866q" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.347891 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpnkx\" (UniqueName: \"kubernetes.io/projected/8fe4d47a-5be0-4130-92e7-da8b26b8940b-kube-api-access-cpnkx\") pod \"ingress-operator-5b745b69d9-bgjpl\" (UID: \"8fe4d47a-5be0-4130-92e7-da8b26b8940b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bgjpl" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.367268 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jx98f\" (UniqueName: \"kubernetes.io/projected/f7c43971-89ed-4b19-8138-0baced4bdd1f-kube-api-access-jx98f\") pod \"service-ca-operator-777779d784-w4vtf\" (UID: \"f7c43971-89ed-4b19-8138-0baced4bdd1f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-w4vtf" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.379964 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-6spsg" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.387330 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320665-rkbh5" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.388939 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2wfp\" (UniqueName: \"kubernetes.io/projected/a9361258-eda9-4ee7-8147-5a720baf1e86-kube-api-access-z2wfp\") pod \"catalog-operator-68c6474976-s72m9\" (UID: \"a9361258-eda9-4ee7-8147-5a720baf1e86\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s72m9" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.390719 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:41 crc kubenswrapper[4840]: E0930 13:58:41.391130 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:41.891113317 +0000 UTC m=+150.520199740 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.394249 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:41 crc kubenswrapper[4840]: E0930 13:58:41.395008 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:41.894981529 +0000 UTC m=+150.524067972 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.429159 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xxjf\" (UniqueName: \"kubernetes.io/projected/641b86bf-c833-495c-9444-0c3c28bac9cc-kube-api-access-4xxjf\") pod \"machine-config-server-zb5zq\" (UID: \"641b86bf-c833-495c-9444-0c3c28bac9cc\") " pod="openshift-machine-config-operator/machine-config-server-zb5zq" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.447733 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-zb5zq" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.448220 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b526l\" (UniqueName: \"kubernetes.io/projected/3869773f-0c18-4f5d-a0e8-2f35ad030b5c-kube-api-access-b526l\") pod \"ingress-canary-8j25v\" (UID: \"3869773f-0c18-4f5d-a0e8-2f35ad030b5c\") " pod="openshift-ingress-canary/ingress-canary-8j25v" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.470784 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfs2k\" (UniqueName: \"kubernetes.io/projected/c4ecdaa2-b6f9-4682-b115-eb12a0060263-kube-api-access-lfs2k\") pod \"csi-hostpathplugin-56p76\" (UID: \"c4ecdaa2-b6f9-4682-b115-eb12a0060263\") " pod="hostpath-provisioner/csi-hostpathplugin-56p76" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.496821 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:41 crc kubenswrapper[4840]: E0930 13:58:41.497350 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:41.997321984 +0000 UTC m=+150.626408437 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:41 crc kubenswrapper[4840]: W0930 13:58:41.513060 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf0515176_8bf5_46a4_a905_7d308f64449e.slice/crio-ecaad802739e8745920d6c23517012c73e7b174e553dde5c346bed2c9b7d3fbb WatchSource:0}: Error finding container ecaad802739e8745920d6c23517012c73e7b174e553dde5c346bed2c9b7d3fbb: Status 404 returned error can't find the container with id ecaad802739e8745920d6c23517012c73e7b174e553dde5c346bed2c9b7d3fbb Sep 30 13:58:41 crc kubenswrapper[4840]: W0930 13:58:41.525526 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4643a53d_3236_4642_9206_603d20a15129.slice/crio-09c61b1bd20b3fde0b590f79b4544b092f9c0d1a2809e86781fd0cf68abd81e1 WatchSource:0}: Error finding container 09c61b1bd20b3fde0b590f79b4544b092f9c0d1a2809e86781fd0cf68abd81e1: Status 404 returned error can't find the container with id 09c61b1bd20b3fde0b590f79b4544b092f9c0d1a2809e86781fd0cf68abd81e1 Sep 30 13:58:41 crc kubenswrapper[4840]: W0930 13:58:41.527234 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8d0002f3_3315_4c27_bc39_2e5989b67d6b.slice/crio-eb03ef0ef76ca920a98c9e5c4439e003eb7c3278d20db930548b308d9de3b781 WatchSource:0}: Error finding container eb03ef0ef76ca920a98c9e5c4439e003eb7c3278d20db930548b308d9de3b781: Status 404 returned error can't find the container with id eb03ef0ef76ca920a98c9e5c4439e003eb7c3278d20db930548b308d9de3b781 Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.584705 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bgjpl" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.598821 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:41 crc kubenswrapper[4840]: E0930 13:58:41.600270 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:42.100250544 +0000 UTC m=+150.729337017 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.624377 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-czbs2" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.654202 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s72m9" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.666341 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-w4vtf" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.701476 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.701759 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-8j25v" Sep 30 13:58:41 crc kubenswrapper[4840]: E0930 13:58:41.702741 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:42.202698982 +0000 UTC m=+150.831785485 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.743303 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-56p76" Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.803817 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:41 crc kubenswrapper[4840]: E0930 13:58:41.804155 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:42.304140883 +0000 UTC m=+150.933227306 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.859802 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-s9lx8" event={"ID":"f0515176-8bf5-46a4-a905-7d308f64449e","Type":"ContainerStarted","Data":"ecaad802739e8745920d6c23517012c73e7b174e553dde5c346bed2c9b7d3fbb"} Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.871199 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftgw2" event={"ID":"2842cd75-1ff5-4a3f-b7a0-8205388f4a17","Type":"ContainerStarted","Data":"39c1272e3385ccb51e5260e86640e906d78ac3858150479e80f9fa91b40c984f"} Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.872234 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9rnrg" event={"ID":"5b44f815-36aa-42d8-9c3e-fb8d956c05c2","Type":"ContainerStarted","Data":"0d71bfdfaf6a3451f9b566ea98eeecf3585d038a41be28abb5c3bfe59c0985a6"} Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.875269 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-kdwqh" event={"ID":"dde0d17e-6bd0-4acb-b8b8-a8b0bc988a6e","Type":"ContainerStarted","Data":"75946e7f410462033b435c09e6baeee583d470d113e36e10754b1c4b7c2e6b19"} Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.877324 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9s79z" event={"ID":"8d0002f3-3315-4c27-bc39-2e5989b67d6b","Type":"ContainerStarted","Data":"eb03ef0ef76ca920a98c9e5c4439e003eb7c3278d20db930548b308d9de3b781"} Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.881587 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p76w8" event={"ID":"e8eb0c45-6539-478d-96ff-1d43d5d9978b","Type":"ContainerStarted","Data":"c09428699400460140f0c1b7904bf80daf415a11bc468f54a4ae95fb172eafa9"} Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.882953 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-dckvx" event={"ID":"4643a53d-3236-4642-9206-603d20a15129","Type":"ContainerStarted","Data":"09c61b1bd20b3fde0b590f79b4544b092f9c0d1a2809e86781fd0cf68abd81e1"} Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.908420 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:41 crc kubenswrapper[4840]: E0930 13:58:41.908572 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:42.408516521 +0000 UTC m=+151.037602954 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.908793 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:41 crc kubenswrapper[4840]: E0930 13:58:41.909182 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:42.409172198 +0000 UTC m=+151.038258691 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:41 crc kubenswrapper[4840]: I0930 13:58:41.925679 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m2bq7"] Sep 30 13:58:42 crc kubenswrapper[4840]: I0930 13:58:42.009773 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:42 crc kubenswrapper[4840]: E0930 13:58:42.009959 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:42.509912671 +0000 UTC m=+151.138999104 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:42 crc kubenswrapper[4840]: I0930 13:58:42.010128 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:42 crc kubenswrapper[4840]: E0930 13:58:42.010378 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:42.510369843 +0000 UTC m=+151.139456266 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:42 crc kubenswrapper[4840]: I0930 13:58:42.111110 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:42 crc kubenswrapper[4840]: E0930 13:58:42.111251 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:42.611226298 +0000 UTC m=+151.240312711 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:42 crc kubenswrapper[4840]: I0930 13:58:42.111639 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:42 crc kubenswrapper[4840]: E0930 13:58:42.112675 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:42.612653916 +0000 UTC m=+151.241740339 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:42 crc kubenswrapper[4840]: I0930 13:58:42.198803 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-lk4jg"] Sep 30 13:58:42 crc kubenswrapper[4840]: I0930 13:58:42.198849 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ncblm"] Sep 30 13:58:42 crc kubenswrapper[4840]: I0930 13:58:42.198870 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-bf4gt"] Sep 30 13:58:42 crc kubenswrapper[4840]: I0930 13:58:42.198886 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-jfnms"] Sep 30 13:58:42 crc kubenswrapper[4840]: I0930 13:58:42.214034 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:42 crc kubenswrapper[4840]: E0930 13:58:42.214206 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:42.714179059 +0000 UTC m=+151.343265482 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:42 crc kubenswrapper[4840]: I0930 13:58:42.214511 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:42 crc kubenswrapper[4840]: E0930 13:58:42.214855 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:42.714845937 +0000 UTC m=+151.343932360 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:42 crc kubenswrapper[4840]: I0930 13:58:42.315031 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:42 crc kubenswrapper[4840]: E0930 13:58:42.315206 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:42.815179628 +0000 UTC m=+151.444266051 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:42 crc kubenswrapper[4840]: I0930 13:58:42.315328 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:42 crc kubenswrapper[4840]: E0930 13:58:42.317356 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:42.817341206 +0000 UTC m=+151.446427629 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:42 crc kubenswrapper[4840]: I0930 13:58:42.416434 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:42 crc kubenswrapper[4840]: E0930 13:58:42.416605 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:42.916574378 +0000 UTC m=+151.545660801 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:42 crc kubenswrapper[4840]: I0930 13:58:42.417049 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:42 crc kubenswrapper[4840]: E0930 13:58:42.417460 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:42.917443911 +0000 UTC m=+151.546530334 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:42 crc kubenswrapper[4840]: I0930 13:58:42.522741 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:42 crc kubenswrapper[4840]: E0930 13:58:42.523056 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:43.023041132 +0000 UTC m=+151.652127555 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:42 crc kubenswrapper[4840]: I0930 13:58:42.624719 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:42 crc kubenswrapper[4840]: E0930 13:58:42.625387 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:43.125371496 +0000 UTC m=+151.754457919 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:42 crc kubenswrapper[4840]: I0930 13:58:42.727649 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:42 crc kubenswrapper[4840]: E0930 13:58:42.729015 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:43.228993305 +0000 UTC m=+151.858079728 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:42 crc kubenswrapper[4840]: I0930 13:58:42.729077 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:42 crc kubenswrapper[4840]: E0930 13:58:42.730081 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:43.230055453 +0000 UTC m=+151.859141876 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:42 crc kubenswrapper[4840]: I0930 13:58:42.834853 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:42 crc kubenswrapper[4840]: E0930 13:58:42.835699 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:43.335678654 +0000 UTC m=+151.964765087 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:42 crc kubenswrapper[4840]: I0930 13:58:42.879209 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vckwz"] Sep 30 13:58:42 crc kubenswrapper[4840]: I0930 13:58:42.886444 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-68gcw"] Sep 30 13:58:42 crc kubenswrapper[4840]: I0930 13:58:42.890944 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m2bq7" event={"ID":"d0cf2421-0839-45da-bf4d-2c26e3137d16","Type":"ContainerStarted","Data":"16f2ed2b0b82196af928f95d6fc178c1ce1cb979a8bcbc0edd3a4a1eab6b4392"} Sep 30 13:58:42 crc kubenswrapper[4840]: I0930 13:58:42.890992 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m2bq7" event={"ID":"d0cf2421-0839-45da-bf4d-2c26e3137d16","Type":"ContainerStarted","Data":"f986d9879646af2b995851d4f39f0ebd2b2bf254aa47bc3b1e9a3b2d65703d72"} Sep 30 13:58:42 crc kubenswrapper[4840]: I0930 13:58:42.901874 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-snznw"] Sep 30 13:58:42 crc kubenswrapper[4840]: I0930 13:58:42.937748 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:42 crc kubenswrapper[4840]: E0930 13:58:42.938684 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:43.438637215 +0000 UTC m=+152.067723638 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:42 crc kubenswrapper[4840]: I0930 13:58:42.945839 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bf4gt" event={"ID":"2b720224-3e65-4443-a056-4e8fd226bcea","Type":"ContainerStarted","Data":"d1a4c65e3f7cdc390f50eb6de7e052ec80e67835f4282862a3ef194ccc3bd4fa"} Sep 30 13:58:42 crc kubenswrapper[4840]: I0930 13:58:42.954371 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4942b"] Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.013383 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s72m9"] Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.022917 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-8h224"] Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.036437 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9sxdn"] Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.036526 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9rnrg" event={"ID":"5b44f815-36aa-42d8-9c3e-fb8d956c05c2","Type":"ContainerStarted","Data":"f68681c97253a93d7ae049002486bfa0bf99763da9040c6bd1533d9094742f5c"} Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.036650 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-dckvx" event={"ID":"4643a53d-3236-4642-9206-603d20a15129","Type":"ContainerStarted","Data":"0e27769395109eef875bb29657778d28bda9bf5f3072863e16b9371d77fa3d58"} Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.036755 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sc7xq"] Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.036863 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpb68"] Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.038926 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-wdjfz" event={"ID":"661771aa-bd40-40fd-8a59-dc3b8156e5c5","Type":"ContainerStarted","Data":"8425bcc99851578d42b5dc397401ae3091ff1c21fda2e7fe3746e0cd8dd51f8c"} Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.039318 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-dksqg"] Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.039337 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.040052 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-wdjfz" Sep 30 13:58:43 crc kubenswrapper[4840]: E0930 13:58:43.040219 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:43.540188489 +0000 UTC m=+152.169274912 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.042203 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9szj8"] Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.042994 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-l47x7"] Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.043882 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-zb5zq" event={"ID":"641b86bf-c833-495c-9444-0c3c28bac9cc","Type":"ContainerStarted","Data":"7e8944a6b0e9a242cb9bd264057bce4c72d21a2773aa22a54ea328f972dc8dc2"} Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.044233 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-ptz4x"] Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.044327 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-zb5zq" event={"ID":"641b86bf-c833-495c-9444-0c3c28bac9cc","Type":"ContainerStarted","Data":"bc3b99d4ade41c39c9e52e83678282ec5880c3ad2ed8a7b97517d257ea942670"} Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.046902 4840 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-wdjfz container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.046995 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-wdjfz" podUID="661771aa-bd40-40fd-8a59-dc3b8156e5c5" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.047460 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" event={"ID":"c1b9a905-662c-4e2a-8066-ee8d75e5d45b","Type":"ContainerStarted","Data":"048ac3b46bf7421fdbedd6180431d06ec5e77076bc085b5d21b543ae66da3ab7"} Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.056240 4840 generic.go:334] "Generic (PLEG): container finished" podID="8d0002f3-3315-4c27-bc39-2e5989b67d6b" containerID="c6d9d1ea786f10a69db43835f05cde1cc6d23edd59c89666288d78485d866dce" exitCode=0 Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.056606 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9s79z" event={"ID":"8d0002f3-3315-4c27-bc39-2e5989b67d6b","Type":"ContainerDied","Data":"c6d9d1ea786f10a69db43835f05cde1cc6d23edd59c89666288d78485d866dce"} Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.058588 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p76w8" event={"ID":"e8eb0c45-6539-478d-96ff-1d43d5d9978b","Type":"ContainerStarted","Data":"cc22a548dffdeb45cfa077ac9901a8ee11bcca76f5f53f461a427166e6d5d411"} Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.066607 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-zbnls" event={"ID":"d3508091-c5c6-4dca-9b32-883426154a6e","Type":"ContainerStarted","Data":"6da45c97e13bff8c8471e3d44cd6e9ceddeceb56af5a77a5e3ed945dd9cb394f"} Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.066669 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-zbnls" event={"ID":"d3508091-c5c6-4dca-9b32-883426154a6e","Type":"ContainerStarted","Data":"d83fe9b0277f783f9b668a40c1e9c3f2539ed4bcd5c4932a0bb91ba1ee7516bc"} Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.068723 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" event={"ID":"8ec02940-3b96-41aa-8624-8338ee544fe1","Type":"ContainerStarted","Data":"f14e817d6fe07e3a0778db5401fe8b920cc40981e0697a7bbce7af0edf6260c6"} Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.070399 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-jfnms" event={"ID":"5103e23b-0d63-4e91-8fd9-afd0b755e619","Type":"ContainerStarted","Data":"3b67c9c480797e04c06b546a98655d035df1e7268d1311f41b7a4d2a622bc49b"} Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.071997 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-jfnms" Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.072091 4840 patch_prober.go:28] interesting pod/console-operator-58897d9998-jfnms container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/readyz\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.072128 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-jfnms" podUID="5103e23b-0d63-4e91-8fd9-afd0b755e619" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.12:8443/readyz\": dial tcp 10.217.0.12:8443: connect: connection refused" Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.078789 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-s9lx8" event={"ID":"f0515176-8bf5-46a4-a905-7d308f64449e","Type":"ContainerStarted","Data":"514ec43eca9ff69fbd38b2de8abea721ff30dda0b5f372a3a77e7a36473b7ca2"} Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.081314 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-qjxnf" event={"ID":"9dd57b46-77fc-4b8a-9fc1-a50084165914","Type":"ContainerStarted","Data":"175520c065fff8b959d46113bca4e5564c481dbced39acbba4f757dc0970dfff"} Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.087350 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-v676n" event={"ID":"bb0b27d1-5cb0-4a2c-95d7-012bc0a2d071","Type":"ContainerStarted","Data":"4ffb9583c74a4b41beb97337e21785b147c2c99b04047db95604c7b7b80d2adb"} Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.087944 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-kdwqh" Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.087983 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftgw2" Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.094946 4840 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-ftgw2 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.095034 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftgw2" podUID="2842cd75-1ff5-4a3f-b7a0-8205388f4a17" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.095148 4840 patch_prober.go:28] interesting pod/downloads-7954f5f757-kdwqh container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.095164 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-kdwqh" podUID="dde0d17e-6bd0-4acb-b8b8-a8b0bc988a6e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.114447 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n2cvx"] Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.119176 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z866q"] Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.122509 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kkmmf"] Sep 30 13:58:43 crc kubenswrapper[4840]: W0930 13:58:43.125682 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod82fad711_fedd_452d_83a5_cf19fbf89815.slice/crio-8ee731f327264f522525e7f71783045fedf823b00813efc8a37f318a1b42b1bb WatchSource:0}: Error finding container 8ee731f327264f522525e7f71783045fedf823b00813efc8a37f318a1b42b1bb: Status 404 returned error can't find the container with id 8ee731f327264f522525e7f71783045fedf823b00813efc8a37f318a1b42b1bb Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.135297 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-bgjpl"] Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.142200 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:43 crc kubenswrapper[4840]: E0930 13:58:43.145890 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:43.645872571 +0000 UTC m=+152.274959084 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.146755 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-czbs2"] Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.155078 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-6spsg"] Sep 30 13:58:43 crc kubenswrapper[4840]: W0930 13:58:43.163428 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0ba2ddf1_a9b7_448a_9e48_cdc9e80f3873.slice/crio-206ddef33a1239163770141cbef0890b476fec000f9dfbccc4e5f35e7c71003c WatchSource:0}: Error finding container 206ddef33a1239163770141cbef0890b476fec000f9dfbccc4e5f35e7c71003c: Status 404 returned error can't find the container with id 206ddef33a1239163770141cbef0890b476fec000f9dfbccc4e5f35e7c71003c Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.172850 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-jslz5"] Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.179085 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-w4vtf"] Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.191745 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-56p76"] Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.213220 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-8j25v"] Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.247243 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:43 crc kubenswrapper[4840]: E0930 13:58:43.247503 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:43.747477536 +0000 UTC m=+152.376563969 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.247747 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:43 crc kubenswrapper[4840]: E0930 13:58:43.248141 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:43.748129393 +0000 UTC m=+152.377215816 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.268006 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320665-rkbh5"] Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.304868 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftgw2" podStartSLOduration=128.304852332 podStartE2EDuration="2m8.304852332s" podCreationTimestamp="2025-09-30 13:56:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:43.304081682 +0000 UTC m=+151.933168125" watchObservedRunningTime="2025-09-30 13:58:43.304852332 +0000 UTC m=+151.933938755" Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.345298 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-s9lx8" podStartSLOduration=128.345277951 podStartE2EDuration="2m8.345277951s" podCreationTimestamp="2025-09-30 13:56:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:43.344399548 +0000 UTC m=+151.973485971" watchObservedRunningTime="2025-09-30 13:58:43.345277951 +0000 UTC m=+151.974364374" Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.348941 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:43 crc kubenswrapper[4840]: E0930 13:58:43.349502 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:43.849484222 +0000 UTC m=+152.478570645 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:43 crc kubenswrapper[4840]: W0930 13:58:43.367074 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod850a2c18_882c_495f_a7e8_743e425a8b61.slice/crio-dc3ae9ae3faf80ae6dcd38bad5d98f623859292dcaa35b5b4151a7518ed73871 WatchSource:0}: Error finding container dc3ae9ae3faf80ae6dcd38bad5d98f623859292dcaa35b5b4151a7518ed73871: Status 404 returned error can't find the container with id dc3ae9ae3faf80ae6dcd38bad5d98f623859292dcaa35b5b4151a7518ed73871 Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.391589 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-jfnms" podStartSLOduration=129.391544164 podStartE2EDuration="2m9.391544164s" podCreationTimestamp="2025-09-30 13:56:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:43.388679568 +0000 UTC m=+152.017766021" watchObservedRunningTime="2025-09-30 13:58:43.391544164 +0000 UTC m=+152.020630607" Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.430718 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-dckvx" podStartSLOduration=128.430700298 podStartE2EDuration="2m8.430700298s" podCreationTimestamp="2025-09-30 13:56:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:43.429078116 +0000 UTC m=+152.058164539" watchObservedRunningTime="2025-09-30 13:58:43.430700298 +0000 UTC m=+152.059786731" Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.450960 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:43 crc kubenswrapper[4840]: E0930 13:58:43.451555 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:43.951533439 +0000 UTC m=+152.580619932 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.463681 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-kdwqh" podStartSLOduration=129.463656629 podStartE2EDuration="2m9.463656629s" podCreationTimestamp="2025-09-30 13:56:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:43.462805537 +0000 UTC m=+152.091891960" watchObservedRunningTime="2025-09-30 13:58:43.463656629 +0000 UTC m=+152.092743072" Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.508137 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-zb5zq" podStartSLOduration=5.5081155939999995 podStartE2EDuration="5.508115594s" podCreationTimestamp="2025-09-30 13:58:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:43.507913189 +0000 UTC m=+152.136999612" watchObservedRunningTime="2025-09-30 13:58:43.508115594 +0000 UTC m=+152.137202017" Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.555624 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:43 crc kubenswrapper[4840]: E0930 13:58:43.555775 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:44.055755363 +0000 UTC m=+152.684841786 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.555955 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:43 crc kubenswrapper[4840]: E0930 13:58:43.556342 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:44.056329229 +0000 UTC m=+152.685415652 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.557041 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-wdjfz" podStartSLOduration=129.557023387 podStartE2EDuration="2m9.557023387s" podCreationTimestamp="2025-09-30 13:56:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:43.555046005 +0000 UTC m=+152.184132438" watchObservedRunningTime="2025-09-30 13:58:43.557023387 +0000 UTC m=+152.186109810" Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.592284 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-zbnls" podStartSLOduration=128.592267608 podStartE2EDuration="2m8.592267608s" podCreationTimestamp="2025-09-30 13:56:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:43.591761335 +0000 UTC m=+152.220847768" watchObservedRunningTime="2025-09-30 13:58:43.592267608 +0000 UTC m=+152.221354031" Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.609958 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.611693 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-s9lx8" Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.632444 4840 patch_prober.go:28] interesting pod/router-default-5444994796-s9lx8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 13:58:43 crc kubenswrapper[4840]: [-]has-synced failed: reason withheld Sep 30 13:58:43 crc kubenswrapper[4840]: [+]process-running ok Sep 30 13:58:43 crc kubenswrapper[4840]: healthz check failed Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.632506 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s9lx8" podUID="f0515176-8bf5-46a4-a905-7d308f64449e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.659294 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:43 crc kubenswrapper[4840]: E0930 13:58:43.660355 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:44.160337887 +0000 UTC m=+152.789424300 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.672828 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-v676n" podStartSLOduration=129.672804347 podStartE2EDuration="2m9.672804347s" podCreationTimestamp="2025-09-30 13:56:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:43.672694064 +0000 UTC m=+152.301780507" watchObservedRunningTime="2025-09-30 13:58:43.672804347 +0000 UTC m=+152.301890770" Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.716168 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-qjxnf" podStartSLOduration=129.716144462 podStartE2EDuration="2m9.716144462s" podCreationTimestamp="2025-09-30 13:56:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:43.715019262 +0000 UTC m=+152.344105685" watchObservedRunningTime="2025-09-30 13:58:43.716144462 +0000 UTC m=+152.345230885" Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.758498 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m2bq7" podStartSLOduration=129.75845907 podStartE2EDuration="2m9.75845907s" podCreationTimestamp="2025-09-30 13:56:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:43.746384481 +0000 UTC m=+152.375470904" watchObservedRunningTime="2025-09-30 13:58:43.75845907 +0000 UTC m=+152.387545493" Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.761399 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:43 crc kubenswrapper[4840]: E0930 13:58:43.761803 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:44.261791848 +0000 UTC m=+152.890878271 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.864483 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:43 crc kubenswrapper[4840]: E0930 13:58:43.864820 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:44.364792221 +0000 UTC m=+152.993878644 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.865333 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:43 crc kubenswrapper[4840]: E0930 13:58:43.865908 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:44.365880639 +0000 UTC m=+152.994967062 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.967608 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:43 crc kubenswrapper[4840]: E0930 13:58:43.967892 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:44.467863095 +0000 UTC m=+153.096949518 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:43 crc kubenswrapper[4840]: I0930 13:58:43.968145 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:43 crc kubenswrapper[4840]: E0930 13:58:43.968469 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:44.46845395 +0000 UTC m=+153.097540373 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.068840 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:44 crc kubenswrapper[4840]: E0930 13:58:44.069248 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:44.569228754 +0000 UTC m=+153.198315177 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.134745 4840 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-snznw container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" start-of-body= Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.134839 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-snznw" podUID="32d505eb-6035-488f-b154-2699e8af36ed" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.142730 4840 generic.go:334] "Generic (PLEG): container finished" podID="2b720224-3e65-4443-a056-4e8fd226bcea" containerID="c55cd8faa0ceb63a5d4bf424cc0012a6936b12b0c33e63ac30169b8cc5a1f42d" exitCode=0 Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.147245 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-snznw" event={"ID":"32d505eb-6035-488f-b154-2699e8af36ed","Type":"ContainerStarted","Data":"f2cea8a3f47b88cffd9635d93bde067d17e1f83cbb3211c684b271f45b73f627"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.147295 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-snznw" event={"ID":"32d505eb-6035-488f-b154-2699e8af36ed","Type":"ContainerStarted","Data":"38a1b3e8457273c6176cf0b8c5674d0517c65eb9f14330018e2ceb47ddc97a14"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.147308 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bf4gt" event={"ID":"2b720224-3e65-4443-a056-4e8fd226bcea","Type":"ContainerDied","Data":"c55cd8faa0ceb63a5d4bf424cc0012a6936b12b0c33e63ac30169b8cc5a1f42d"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.147328 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-snznw" Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.148532 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-snznw" podStartSLOduration=129.148517949 podStartE2EDuration="2m9.148517949s" podCreationTimestamp="2025-09-30 13:56:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:44.146998289 +0000 UTC m=+152.776084712" watchObservedRunningTime="2025-09-30 13:58:44.148517949 +0000 UTC m=+152.777604372" Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.148683 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-jslz5" event={"ID":"1454044a-1464-42f3-92c4-f6093e3bd7d1","Type":"ContainerStarted","Data":"35df7c82a980d4a2531f06d3946e39897ac7b8036dd43886f8307bc1de1493fe"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.160801 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9rnrg" event={"ID":"5b44f815-36aa-42d8-9c3e-fb8d956c05c2","Type":"ContainerStarted","Data":"796643fe4eca88eff00d9133a4aca5e3773bd99c2f3a81746a7a85fc8e3213d5"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.166902 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4942b" event={"ID":"0ba2ddf1-a9b7-448a-9e48-cdc9e80f3873","Type":"ContainerStarted","Data":"206ddef33a1239163770141cbef0890b476fec000f9dfbccc4e5f35e7c71003c"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.169392 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9sxdn" event={"ID":"69f4620c-5949-4527-aecb-9765f22e653e","Type":"ContainerStarted","Data":"436453ad22ad05545fb2c15872573cc82e0cca2d8c1723462193d6b6e3029da3"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.170348 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:44 crc kubenswrapper[4840]: E0930 13:58:44.170675 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:44.670662944 +0000 UTC m=+153.299749367 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.176052 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p76w8" event={"ID":"e8eb0c45-6539-478d-96ff-1d43d5d9978b","Type":"ContainerStarted","Data":"48c91c234c10a550994a1e6cef801e38228504a8f6cccb266a5394699f4341ef"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.182925 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sc7xq" event={"ID":"ea0a1a12-942c-4ca3-b649-5ad87af478c8","Type":"ContainerStarted","Data":"c679f19c89f98ca14a286670f5a36107842c3e7973ff5c1ea3be7bc9b6210906"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.182985 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sc7xq" event={"ID":"ea0a1a12-942c-4ca3-b649-5ad87af478c8","Type":"ContainerStarted","Data":"6b56d95e1f8c7ccdf00c7ecf17507fb5e42ec1f95ad6f97abea8f6ec9325a4a3"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.196858 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p76w8" podStartSLOduration=130.196833196 podStartE2EDuration="2m10.196833196s" podCreationTimestamp="2025-09-30 13:56:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:44.195226073 +0000 UTC m=+152.824312496" watchObservedRunningTime="2025-09-30 13:58:44.196833196 +0000 UTC m=+152.825919619" Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.202014 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-68gcw" event={"ID":"94d7d8b4-fc87-48bc-ac61-84e39213d89a","Type":"ContainerStarted","Data":"834279422b4e1de4a93a1c356622dfc74cc43e30b3ab922fc0531dfdf807a3f7"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.202065 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-68gcw" event={"ID":"94d7d8b4-fc87-48bc-ac61-84e39213d89a","Type":"ContainerStarted","Data":"8642ed350d3485bf2c243ecfe44d24f79b34daf6915e2e31de1d8f6379e557e5"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.206514 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-jfnms" event={"ID":"5103e23b-0d63-4e91-8fd9-afd0b755e619","Type":"ContainerStarted","Data":"afe275fdae5b602a027aa2c364dee944f8c9906c55e180e4923d54a67c8320d2"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.207399 4840 patch_prober.go:28] interesting pod/console-operator-58897d9998-jfnms container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/readyz\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.207449 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-jfnms" podUID="5103e23b-0d63-4e91-8fd9-afd0b755e619" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.12:8443/readyz\": dial tcp 10.217.0.12:8443: connect: connection refused" Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.208917 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n2cvx" event={"ID":"f5ac3e36-a48a-477a-b83e-6d9acaba6169","Type":"ContainerStarted","Data":"6d26ac19e6755dcb0bff5ddeae2850a716f981a0fcb2f7d05db61d091db19f71"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.217011 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9rnrg" podStartSLOduration=130.216987728 podStartE2EDuration="2m10.216987728s" podCreationTimestamp="2025-09-30 13:56:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:44.210737123 +0000 UTC m=+152.839823546" watchObservedRunningTime="2025-09-30 13:58:44.216987728 +0000 UTC m=+152.846074151" Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.218277 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-czbs2" event={"ID":"7f88fdb6-dac3-4508-95a6-976cd084c4a7","Type":"ContainerStarted","Data":"3f19ff04133b8f084bf4a96cf69cb947d5680272f01484c6b21b81ab6fa1036e"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.218391 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-czbs2" event={"ID":"7f88fdb6-dac3-4508-95a6-976cd084c4a7","Type":"ContainerStarted","Data":"d190f6f01c69c968ada5ece6b570fcd72d69f475c03b86fa55eb36db79d49f89"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.218463 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-czbs2" Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.220682 4840 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-czbs2 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.220742 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-czbs2" podUID="7f88fdb6-dac3-4508-95a6-976cd084c4a7" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.231036 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kkmmf" event={"ID":"9955a9fc-27dc-4765-8bb9-e84fbb5448f1","Type":"ContainerStarted","Data":"f88de2719f5d81c22d4228af68094034ba82967314255d152481a3f331f092ee"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.232283 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sc7xq" podStartSLOduration=130.232263792 podStartE2EDuration="2m10.232263792s" podCreationTimestamp="2025-09-30 13:56:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:44.231182953 +0000 UTC m=+152.860269376" watchObservedRunningTime="2025-09-30 13:58:44.232263792 +0000 UTC m=+152.861350215" Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.260596 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-czbs2" podStartSLOduration=129.26057667 podStartE2EDuration="2m9.26057667s" podCreationTimestamp="2025-09-30 13:56:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:44.258137146 +0000 UTC m=+152.887223589" watchObservedRunningTime="2025-09-30 13:58:44.26057667 +0000 UTC m=+152.889663093" Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.264968 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-8j25v" event={"ID":"3869773f-0c18-4f5d-a0e8-2f35ad030b5c","Type":"ContainerStarted","Data":"27d77e501fdd88cd3bd8dce7c14823933fab60a6b6c5bdcb995ebea64d103cd6"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.276394 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:44 crc kubenswrapper[4840]: E0930 13:58:44.276641 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:44.776624404 +0000 UTC m=+153.405710827 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.276848 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.277203 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kkmmf" podStartSLOduration=129.277185209 podStartE2EDuration="2m9.277185209s" podCreationTimestamp="2025-09-30 13:56:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:44.276544922 +0000 UTC m=+152.905631345" watchObservedRunningTime="2025-09-30 13:58:44.277185209 +0000 UTC m=+152.906271632" Sep 30 13:58:44 crc kubenswrapper[4840]: E0930 13:58:44.278765 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:44.778747461 +0000 UTC m=+153.407833884 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.294889 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9s79z" event={"ID":"8d0002f3-3315-4c27-bc39-2e5989b67d6b","Type":"ContainerStarted","Data":"8aa86d71a3e9f8fc5e2699bbf57daac8f4ca2cb2e2c1abd30d663cf1d2c02172"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.295839 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9s79z" Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.299744 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z866q" event={"ID":"eb59a261-a39c-4be5-86f0-a14643aef960","Type":"ContainerStarted","Data":"51501f5d1bd727dbc1e68050c2bf9556e12c927516027512e5c12bf6af83bb8c"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.305895 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vckwz" event={"ID":"46efdeea-a4d2-4ff5-898d-9ee516105cf4","Type":"ContainerStarted","Data":"0d6c1e1a7aa71d062ffe33c3f0c59114275d9e576222bb85a114247fef46ca0e"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.305938 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vckwz" event={"ID":"46efdeea-a4d2-4ff5-898d-9ee516105cf4","Type":"ContainerStarted","Data":"563292efd27afd05dabdca9ce29c8a19e52d9fe8154cd4ad382308bef464a58d"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.325713 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9s79z" podStartSLOduration=130.325689881 podStartE2EDuration="2m10.325689881s" podCreationTimestamp="2025-09-30 13:56:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:44.324812588 +0000 UTC m=+152.953899031" watchObservedRunningTime="2025-09-30 13:58:44.325689881 +0000 UTC m=+152.954776304" Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.327617 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9szj8" event={"ID":"f71cd461-05df-4f5d-bf11-edf4d313996b","Type":"ContainerStarted","Data":"8b7af90e9b5f7373144f9beeccdb366038130ba2f56822caa042cf430fad581b"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.330307 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-ptz4x" event={"ID":"713dae78-c3ff-40b4-a018-f801cf0562dd","Type":"ContainerStarted","Data":"42003b852fe7a2f33adbfc161220a8e6aa2ead9c85fe95bb2bac0fa0bd9fa7e9"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.333671 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-56p76" event={"ID":"c4ecdaa2-b6f9-4682-b115-eb12a0060263","Type":"ContainerStarted","Data":"c47b7417330f00c89129eb871a9b8e0c6a161c84f1fdfd60056d26f8154bf098"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.344874 4840 generic.go:334] "Generic (PLEG): container finished" podID="8ec02940-3b96-41aa-8624-8338ee544fe1" containerID="4550d62a43952c8155ffb002c5d58d0a2f2ee5c133fc0beb44aec871a6dc257c" exitCode=0 Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.348085 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" event={"ID":"8ec02940-3b96-41aa-8624-8338ee544fe1","Type":"ContainerDied","Data":"4550d62a43952c8155ffb002c5d58d0a2f2ee5c133fc0beb44aec871a6dc257c"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.363467 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8h224" event={"ID":"82fad711-fedd-452d-83a5-cf19fbf89815","Type":"ContainerStarted","Data":"8ee731f327264f522525e7f71783045fedf823b00813efc8a37f318a1b42b1bb"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.366826 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dksqg" event={"ID":"771e14ef-523d-4eab-b13b-87a112fa11bf","Type":"ContainerStarted","Data":"1490f90c21b49f8bbc32623081a0b033de8a85ea44b7ce36dcd83bcfd0141398"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.366864 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dksqg" event={"ID":"771e14ef-523d-4eab-b13b-87a112fa11bf","Type":"ContainerStarted","Data":"266eb2ac823897c8b7ada3a6e647ebccaf3bab6d4d748e461cafd54b38411e5e"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.378781 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.379132 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vckwz" podStartSLOduration=129.379110013 podStartE2EDuration="2m9.379110013s" podCreationTimestamp="2025-09-30 13:56:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:44.347724293 +0000 UTC m=+152.976810716" watchObservedRunningTime="2025-09-30 13:58:44.379110013 +0000 UTC m=+153.008196436" Sep 30 13:58:44 crc kubenswrapper[4840]: E0930 13:58:44.379530 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:44.879511484 +0000 UTC m=+153.508597907 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.396462 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s72m9" event={"ID":"a9361258-eda9-4ee7-8147-5a720baf1e86","Type":"ContainerStarted","Data":"57ac385cec86fcc25c43b6edf12c54ae1dbd4e572eebc429631701a167565f99"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.396511 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s72m9" event={"ID":"a9361258-eda9-4ee7-8147-5a720baf1e86","Type":"ContainerStarted","Data":"95ada3ee1415fc8e5d9d214cf80c9ca32b72709508a26e5243867d1f2f938324"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.396988 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s72m9" Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.399385 4840 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-s72m9 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" start-of-body= Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.399443 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s72m9" podUID="a9361258-eda9-4ee7-8147-5a720baf1e86" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.413881 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-w4vtf" event={"ID":"f7c43971-89ed-4b19-8138-0baced4bdd1f","Type":"ContainerStarted","Data":"e0e3e943f85999471ae3eec6a7053a5214c821184f7317a5f838fc6cd0deace2"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.413924 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-w4vtf" event={"ID":"f7c43971-89ed-4b19-8138-0baced4bdd1f","Type":"ContainerStarted","Data":"ad13595776d27ae8b4843d1f49bf26d88eb81e9237841e6b922d2751c43111ce"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.434102 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s72m9" podStartSLOduration=129.434082536 podStartE2EDuration="2m9.434082536s" podCreationTimestamp="2025-09-30 13:56:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:44.432817852 +0000 UTC m=+153.061904275" watchObservedRunningTime="2025-09-30 13:58:44.434082536 +0000 UTC m=+153.063168959" Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.436737 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-6spsg" event={"ID":"f06465db-0dee-4f91-9e13-e46810c0e20e","Type":"ContainerStarted","Data":"11a15018e4a69db6a29736b96eaa141908290a0056d574ba1002b1ae17b4864e"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.436804 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-6spsg" event={"ID":"f06465db-0dee-4f91-9e13-e46810c0e20e","Type":"ContainerStarted","Data":"65bbbc90d19b6bb590d9dd6a5cb88c0da2ce89f67798df8579c505c96e7496a7"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.456616 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-w4vtf" podStartSLOduration=129.456594441 podStartE2EDuration="2m9.456594441s" podCreationTimestamp="2025-09-30 13:56:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:44.455856641 +0000 UTC m=+153.084943064" watchObservedRunningTime="2025-09-30 13:58:44.456594441 +0000 UTC m=+153.085680864" Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.461149 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bgjpl" event={"ID":"8fe4d47a-5be0-4130-92e7-da8b26b8940b","Type":"ContainerStarted","Data":"9dac48867e709335fde5132c54854dd8f145be371ec7d6b03f52974c0e030732"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.477514 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-6spsg" podStartSLOduration=129.477495233 podStartE2EDuration="2m9.477495233s" podCreationTimestamp="2025-09-30 13:56:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:44.477249427 +0000 UTC m=+153.106335850" watchObservedRunningTime="2025-09-30 13:58:44.477495233 +0000 UTC m=+153.106581646" Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.485387 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:44 crc kubenswrapper[4840]: E0930 13:58:44.486383 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:44.986371448 +0000 UTC m=+153.615457871 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.489946 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" event={"ID":"c1b9a905-662c-4e2a-8066-ee8d75e5d45b","Type":"ContainerStarted","Data":"a8aca6442d2827c689d94cb212fdbd982789b44b0113a53ff16a3cb3e0364162"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.492441 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.497492 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-l47x7" event={"ID":"b4d78aa2-9ff6-404d-9b69-b8da8dfb2f9f","Type":"ContainerStarted","Data":"2d49062ab33a1f04bec1ad95e05a18ab1ad6bcff00f17855eb8c7873dd92dafb"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.497569 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-l47x7" event={"ID":"b4d78aa2-9ff6-404d-9b69-b8da8dfb2f9f","Type":"ContainerStarted","Data":"6aa837ee116cc34b08f81708e0fea8059698b39efac98c0f5e0c796b020c1554"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.502846 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320665-rkbh5" event={"ID":"850a2c18-882c-495f-a7e8-743e425a8b61","Type":"ContainerStarted","Data":"dc3ae9ae3faf80ae6dcd38bad5d98f623859292dcaa35b5b4151a7518ed73871"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.513471 4840 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-ncblm container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.23:6443/healthz\": dial tcp 10.217.0.23:6443: connect: connection refused" start-of-body= Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.513851 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" podUID="c1b9a905-662c-4e2a-8066-ee8d75e5d45b" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.23:6443/healthz\": dial tcp 10.217.0.23:6443: connect: connection refused" Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.516034 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpb68" event={"ID":"11f26882-b509-4819-9fbb-2c69c062c951","Type":"ContainerStarted","Data":"65cb1ddaeb9c4e2c7f087d8c7a360fd6677c0c582a52986beef33c6fa2e8f535"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.516086 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpb68" event={"ID":"11f26882-b509-4819-9fbb-2c69c062c951","Type":"ContainerStarted","Data":"f0275bdcaa54e3a59479f79c4f644dcb00ba8aab3300dcc611c5e4d11f025e38"} Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.518334 4840 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-wdjfz container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.518371 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-wdjfz" podUID="661771aa-bd40-40fd-8a59-dc3b8156e5c5" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.520053 4840 patch_prober.go:28] interesting pod/downloads-7954f5f757-kdwqh container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.520171 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-kdwqh" podUID="dde0d17e-6bd0-4acb-b8b8-a8b0bc988a6e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.525549 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftgw2" Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.529789 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" podStartSLOduration=130.529764704 podStartE2EDuration="2m10.529764704s" podCreationTimestamp="2025-09-30 13:56:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:44.527059863 +0000 UTC m=+153.156146306" watchObservedRunningTime="2025-09-30 13:58:44.529764704 +0000 UTC m=+153.158851147" Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.589035 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:44 crc kubenswrapper[4840]: E0930 13:58:44.589417 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:45.08938528 +0000 UTC m=+153.718471703 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.589785 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:44 crc kubenswrapper[4840]: E0930 13:58:44.598944 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:45.098925092 +0000 UTC m=+153.728011585 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.641556 4840 patch_prober.go:28] interesting pod/router-default-5444994796-s9lx8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 13:58:44 crc kubenswrapper[4840]: [-]has-synced failed: reason withheld Sep 30 13:58:44 crc kubenswrapper[4840]: [+]process-running ok Sep 30 13:58:44 crc kubenswrapper[4840]: healthz check failed Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.642032 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s9lx8" podUID="f0515176-8bf5-46a4-a905-7d308f64449e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.699248 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:44 crc kubenswrapper[4840]: E0930 13:58:44.699682 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:45.199663244 +0000 UTC m=+153.828749667 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.801918 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:44 crc kubenswrapper[4840]: E0930 13:58:44.802994 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:45.302633306 +0000 UTC m=+153.931719729 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:44 crc kubenswrapper[4840]: I0930 13:58:44.903240 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:44 crc kubenswrapper[4840]: E0930 13:58:44.903959 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:45.403943183 +0000 UTC m=+154.033029596 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.004717 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:45 crc kubenswrapper[4840]: E0930 13:58:45.005077 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:45.505062996 +0000 UTC m=+154.134149409 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.105967 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:45 crc kubenswrapper[4840]: E0930 13:58:45.106145 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:45.606120436 +0000 UTC m=+154.235206859 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.106375 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:45 crc kubenswrapper[4840]: E0930 13:58:45.106717 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:45.606707732 +0000 UTC m=+154.235794225 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.207044 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:45 crc kubenswrapper[4840]: E0930 13:58:45.207247 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:45.707219478 +0000 UTC m=+154.336305901 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.207699 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:45 crc kubenswrapper[4840]: E0930 13:58:45.208066 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:45.70805202 +0000 UTC m=+154.337138453 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.308993 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:45 crc kubenswrapper[4840]: E0930 13:58:45.309204 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:45.809174343 +0000 UTC m=+154.438260766 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.309317 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:45 crc kubenswrapper[4840]: E0930 13:58:45.309706 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:45.809698416 +0000 UTC m=+154.438784849 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.410244 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:45 crc kubenswrapper[4840]: E0930 13:58:45.410461 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:45.910431019 +0000 UTC m=+154.539517452 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.410597 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:45 crc kubenswrapper[4840]: E0930 13:58:45.411105 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:45.911090696 +0000 UTC m=+154.540177119 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.511496 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:45 crc kubenswrapper[4840]: E0930 13:58:45.511706 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:46.011679394 +0000 UTC m=+154.640765817 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.511815 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:45 crc kubenswrapper[4840]: E0930 13:58:45.512177 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:46.012162237 +0000 UTC m=+154.641248660 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.524182 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kkmmf" event={"ID":"9955a9fc-27dc-4765-8bb9-e84fbb5448f1","Type":"ContainerStarted","Data":"506860742f97b2ee512a17149b7853428b9e2d380de786eeb6fea973334a5979"} Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.526759 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9sxdn" event={"ID":"69f4620c-5949-4527-aecb-9765f22e653e","Type":"ContainerStarted","Data":"197573a760764be4cd5a54b4ab1482135b75fdde7cb901962ab8ecb25bd6e25d"} Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.526809 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9sxdn" event={"ID":"69f4620c-5949-4527-aecb-9765f22e653e","Type":"ContainerStarted","Data":"91c5aad07bd3beeeb284b5d2594cd473f58fe86cfff5734dddfabe596c9a54f8"} Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.526873 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9sxdn" Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.528746 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dksqg" event={"ID":"771e14ef-523d-4eab-b13b-87a112fa11bf","Type":"ContainerStarted","Data":"82ef7c2140e8f615e50d4dbd1db96c480d6265f1c30cff3582d0005188f539f3"} Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.531235 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z866q" event={"ID":"eb59a261-a39c-4be5-86f0-a14643aef960","Type":"ContainerStarted","Data":"60006826c309233a4ea758d98ddb5454f1dce5c814ebd448c6f5142054f42cce"} Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.533790 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bgjpl" event={"ID":"8fe4d47a-5be0-4130-92e7-da8b26b8940b","Type":"ContainerStarted","Data":"9fe51e0e9cd3d69b37220d396f1b12e3368a560cfdd78f52260112ee7e680d90"} Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.533836 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bgjpl" event={"ID":"8fe4d47a-5be0-4130-92e7-da8b26b8940b","Type":"ContainerStarted","Data":"84efb404982e218c586b616a3e756dd0ff56e1f99f4309e398e0b6b7bac78e8f"} Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.538459 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9szj8" event={"ID":"f71cd461-05df-4f5d-bf11-edf4d313996b","Type":"ContainerStarted","Data":"34b25d59e37f0e717f815cdd1bb7a7d1e7afc7e85c3bea71ba7302e19007371f"} Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.544390 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bf4gt" event={"ID":"2b720224-3e65-4443-a056-4e8fd226bcea","Type":"ContainerStarted","Data":"bf22431e06a42b759b3ede5b4f321beb52b045b006b63688bb6fdff8851ace02"} Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.549943 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8h224" event={"ID":"82fad711-fedd-452d-83a5-cf19fbf89815","Type":"ContainerStarted","Data":"b6d7c1b8a4f13ce256bf9a2efaeffb5d7f90a9feaaa4614101588ca6029ea68a"} Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.549999 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8h224" event={"ID":"82fad711-fedd-452d-83a5-cf19fbf89815","Type":"ContainerStarted","Data":"0b050adfceafe9fae258b1c55a22cd05370023ae493336dc5dfc9872be3c73ba"} Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.558075 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-l47x7" event={"ID":"b4d78aa2-9ff6-404d-9b69-b8da8dfb2f9f","Type":"ContainerStarted","Data":"ead83c12f88489cd3bbf8f7f7d030559fd9dc080f50702e3b96720177315c767"} Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.560029 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpb68" podStartSLOduration=131.560018102 podStartE2EDuration="2m11.560018102s" podCreationTimestamp="2025-09-30 13:56:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:44.59620136 +0000 UTC m=+153.225287783" watchObservedRunningTime="2025-09-30 13:58:45.560018102 +0000 UTC m=+154.189104525" Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.560934 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9sxdn" podStartSLOduration=130.560929626 podStartE2EDuration="2m10.560929626s" podCreationTimestamp="2025-09-30 13:56:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:45.559889999 +0000 UTC m=+154.188976422" watchObservedRunningTime="2025-09-30 13:58:45.560929626 +0000 UTC m=+154.190016049" Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.573271 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320665-rkbh5" event={"ID":"850a2c18-882c-495f-a7e8-743e425a8b61","Type":"ContainerStarted","Data":"bc5d3b48314ba858d8d83c52b612834e9bda381052fd43502a08ffde2724327d"} Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.582924 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-68gcw" event={"ID":"94d7d8b4-fc87-48bc-ac61-84e39213d89a","Type":"ContainerStarted","Data":"9c9ff1408e2cc23258c722c9d56d1a779af6d2280ba24080a5c2c8e62c46519b"} Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.583177 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-68gcw" Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.587469 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" event={"ID":"8ec02940-3b96-41aa-8624-8338ee544fe1","Type":"ContainerStarted","Data":"abe52c0f0a13705d57ab8165ddea46a139430d60720499a8742ad523477ba6f0"} Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.587542 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" event={"ID":"8ec02940-3b96-41aa-8624-8338ee544fe1","Type":"ContainerStarted","Data":"cf652a29b6a01d47d53207f84e562abbe47e95281197a05189b28bb7cae04297"} Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.590906 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z866q" podStartSLOduration=130.590890888 podStartE2EDuration="2m10.590890888s" podCreationTimestamp="2025-09-30 13:56:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:45.590040665 +0000 UTC m=+154.219127088" watchObservedRunningTime="2025-09-30 13:58:45.590890888 +0000 UTC m=+154.219977311" Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.591213 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-ptz4x" event={"ID":"713dae78-c3ff-40b4-a018-f801cf0562dd","Type":"ContainerStarted","Data":"c784c43c742c0d8fd028705ffd9e660b77c55f467df2cc8a589ebc14e39fd653"} Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.591260 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-ptz4x" event={"ID":"713dae78-c3ff-40b4-a018-f801cf0562dd","Type":"ContainerStarted","Data":"ff6fca149da4c50786ff5c6c9e1121897bc6f6d6ca8b917eded4f27dcb00dade"} Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.596994 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bf4gt" Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.597027 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-jslz5" event={"ID":"1454044a-1464-42f3-92c4-f6093e3bd7d1","Type":"ContainerStarted","Data":"a31e13e780747bc910e6d021b0e98f8dd1be7e2d76bbd8c89e9d807af180fa8e"} Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.597044 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-jslz5" event={"ID":"1454044a-1464-42f3-92c4-f6093e3bd7d1","Type":"ContainerStarted","Data":"5cf6ad9ebf32ca9e28f20a3ad66c4871bb441856d5817d65620d4b639c51d64e"} Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.597074 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bf4gt" Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.598593 4840 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-bf4gt container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.24:8443/livez\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.598637 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bf4gt" podUID="2b720224-3e65-4443-a056-4e8fd226bcea" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.24:8443/livez\": dial tcp 10.217.0.24:8443: connect: connection refused" Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.604957 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n2cvx" event={"ID":"f5ac3e36-a48a-477a-b83e-6d9acaba6169","Type":"ContainerStarted","Data":"fe297a08c2e1dd3acd3497a73012ea885afe6bbb0323f922a8ad0b12bd28ca92"} Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.605707 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n2cvx" Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.613886 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:45 crc kubenswrapper[4840]: E0930 13:58:45.615983 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:46.11595885 +0000 UTC m=+154.745045343 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.621894 4840 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-n2cvx container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:5443/healthz\": dial tcp 10.217.0.32:5443: connect: connection refused" start-of-body= Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.622081 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n2cvx" podUID="f5ac3e36-a48a-477a-b83e-6d9acaba6169" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.32:5443/healthz\": dial tcp 10.217.0.32:5443: connect: connection refused" Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.624494 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-8j25v" event={"ID":"3869773f-0c18-4f5d-a0e8-2f35ad030b5c","Type":"ContainerStarted","Data":"dce26adbb7d52c3d9813a032d86f362dab491b66e4bba340115698dacd45fe67"} Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.632232 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4942b" event={"ID":"0ba2ddf1-a9b7-448a-9e48-cdc9e80f3873","Type":"ContainerStarted","Data":"d1865a79873a5c98f8f3b9382b3798ba25604b1321a0abcdb239e2b7e88179d9"} Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.633375 4840 patch_prober.go:28] interesting pod/router-default-5444994796-s9lx8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 13:58:45 crc kubenswrapper[4840]: [-]has-synced failed: reason withheld Sep 30 13:58:45 crc kubenswrapper[4840]: [+]process-running ok Sep 30 13:58:45 crc kubenswrapper[4840]: healthz check failed Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.636329 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s9lx8" podUID="f0515176-8bf5-46a4-a905-7d308f64449e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.642429 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-l47x7" podStartSLOduration=131.642404859 podStartE2EDuration="2m11.642404859s" podCreationTimestamp="2025-09-30 13:56:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:45.631717577 +0000 UTC m=+154.260804000" watchObservedRunningTime="2025-09-30 13:58:45.642404859 +0000 UTC m=+154.271491282" Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.655896 4840 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-s72m9 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" start-of-body= Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.655941 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s72m9" podUID="a9361258-eda9-4ee7-8147-5a720baf1e86" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.656451 4840 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-czbs2 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.656481 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-czbs2" podUID="7f88fdb6-dac3-4508-95a6-976cd084c4a7" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.657395 4840 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-snznw container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" start-of-body= Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.657422 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-snznw" podUID="32d505eb-6035-488f-b154-2699e8af36ed" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.677928 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dksqg" podStartSLOduration=130.677903067 podStartE2EDuration="2m10.677903067s" podCreationTimestamp="2025-09-30 13:56:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:45.664426711 +0000 UTC m=+154.293513134" watchObservedRunningTime="2025-09-30 13:58:45.677903067 +0000 UTC m=+154.306989490" Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.718883 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:45 crc kubenswrapper[4840]: E0930 13:58:45.723612 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:46.223596935 +0000 UTC m=+154.852683358 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.752463 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8h224" podStartSLOduration=130.752445247 podStartE2EDuration="2m10.752445247s" podCreationTimestamp="2025-09-30 13:56:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:45.750852005 +0000 UTC m=+154.379938448" watchObservedRunningTime="2025-09-30 13:58:45.752445247 +0000 UTC m=+154.381531670" Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.753305 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9szj8" podStartSLOduration=130.75329891 podStartE2EDuration="2m10.75329891s" podCreationTimestamp="2025-09-30 13:56:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:45.712770659 +0000 UTC m=+154.341857082" watchObservedRunningTime="2025-09-30 13:58:45.75329891 +0000 UTC m=+154.382385333" Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.819948 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:45 crc kubenswrapper[4840]: E0930 13:58:45.820409 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:46.320390633 +0000 UTC m=+154.949477056 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.836804 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bgjpl" podStartSLOduration=130.836788826 podStartE2EDuration="2m10.836788826s" podCreationTimestamp="2025-09-30 13:56:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:45.799022918 +0000 UTC m=+154.428109341" watchObservedRunningTime="2025-09-30 13:58:45.836788826 +0000 UTC m=+154.465875249" Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.871605 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-ptz4x" podStartSLOduration=130.871552685 podStartE2EDuration="2m10.871552685s" podCreationTimestamp="2025-09-30 13:56:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:45.865860715 +0000 UTC m=+154.494947138" watchObservedRunningTime="2025-09-30 13:58:45.871552685 +0000 UTC m=+154.500639108" Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.882618 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bf4gt" podStartSLOduration=130.882594717 podStartE2EDuration="2m10.882594717s" podCreationTimestamp="2025-09-30 13:56:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:45.837320711 +0000 UTC m=+154.466407134" watchObservedRunningTime="2025-09-30 13:58:45.882594717 +0000 UTC m=+154.511681140" Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.919720 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-jslz5" podStartSLOduration=130.919702718 podStartE2EDuration="2m10.919702718s" podCreationTimestamp="2025-09-30 13:56:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:45.906272753 +0000 UTC m=+154.535359176" watchObservedRunningTime="2025-09-30 13:58:45.919702718 +0000 UTC m=+154.548789141" Sep 30 13:58:45 crc kubenswrapper[4840]: I0930 13:58:45.921856 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:45 crc kubenswrapper[4840]: E0930 13:58:45.922258 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:46.422247225 +0000 UTC m=+155.051333648 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:46 crc kubenswrapper[4840]: I0930 13:58:46.001285 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29320665-rkbh5" podStartSLOduration=131.001268093 podStartE2EDuration="2m11.001268093s" podCreationTimestamp="2025-09-30 13:56:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:45.980882035 +0000 UTC m=+154.609968448" watchObservedRunningTime="2025-09-30 13:58:46.001268093 +0000 UTC m=+154.630354516" Sep 30 13:58:46 crc kubenswrapper[4840]: I0930 13:58:46.025126 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:46 crc kubenswrapper[4840]: E0930 13:58:46.025711 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:46.525693039 +0000 UTC m=+155.154779462 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:46 crc kubenswrapper[4840]: I0930 13:58:46.066312 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 13:58:46 crc kubenswrapper[4840]: I0930 13:58:46.071715 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-8j25v" podStartSLOduration=8.071694825 podStartE2EDuration="8.071694825s" podCreationTimestamp="2025-09-30 13:58:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:46.068215343 +0000 UTC m=+154.697301766" watchObservedRunningTime="2025-09-30 13:58:46.071694825 +0000 UTC m=+154.700781248" Sep 30 13:58:46 crc kubenswrapper[4840]: I0930 13:58:46.128104 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:46 crc kubenswrapper[4840]: E0930 13:58:46.128444 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:46.628431114 +0000 UTC m=+155.257517537 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:46 crc kubenswrapper[4840]: I0930 13:58:46.230735 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:46 crc kubenswrapper[4840]: E0930 13:58:46.230836 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:46.73081701 +0000 UTC m=+155.359903433 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:46 crc kubenswrapper[4840]: I0930 13:58:46.231296 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:46 crc kubenswrapper[4840]: E0930 13:58:46.231643 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:46.731634872 +0000 UTC m=+155.360721295 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:46 crc kubenswrapper[4840]: I0930 13:58:46.232050 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-68gcw" podStartSLOduration=8.232031562 podStartE2EDuration="8.232031562s" podCreationTimestamp="2025-09-30 13:58:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:46.230628665 +0000 UTC m=+154.859715088" watchObservedRunningTime="2025-09-30 13:58:46.232031562 +0000 UTC m=+154.861117985" Sep 30 13:58:46 crc kubenswrapper[4840]: I0930 13:58:46.329001 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n2cvx" podStartSLOduration=131.328981824 podStartE2EDuration="2m11.328981824s" podCreationTimestamp="2025-09-30 13:56:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:46.320245793 +0000 UTC m=+154.949332216" watchObservedRunningTime="2025-09-30 13:58:46.328981824 +0000 UTC m=+154.958068247" Sep 30 13:58:46 crc kubenswrapper[4840]: I0930 13:58:46.329142 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4942b" podStartSLOduration=131.329139648 podStartE2EDuration="2m11.329139648s" podCreationTimestamp="2025-09-30 13:56:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:46.267577331 +0000 UTC m=+154.896663754" watchObservedRunningTime="2025-09-30 13:58:46.329139648 +0000 UTC m=+154.958226071" Sep 30 13:58:46 crc kubenswrapper[4840]: I0930 13:58:46.332392 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:46 crc kubenswrapper[4840]: E0930 13:58:46.332654 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:46.83262388 +0000 UTC m=+155.461710303 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:46 crc kubenswrapper[4840]: I0930 13:58:46.333002 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:46 crc kubenswrapper[4840]: E0930 13:58:46.333374 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:46.83335875 +0000 UTC m=+155.462445173 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:46 crc kubenswrapper[4840]: I0930 13:58:46.373287 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" podStartSLOduration=132.373272125 podStartE2EDuration="2m12.373272125s" podCreationTimestamp="2025-09-30 13:56:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:46.371822146 +0000 UTC m=+155.000908559" watchObservedRunningTime="2025-09-30 13:58:46.373272125 +0000 UTC m=+155.002358548" Sep 30 13:58:46 crc kubenswrapper[4840]: I0930 13:58:46.434065 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:46 crc kubenswrapper[4840]: E0930 13:58:46.434436 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:46.93441154 +0000 UTC m=+155.563497963 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:46 crc kubenswrapper[4840]: I0930 13:58:46.482141 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-jfnms" Sep 30 13:58:46 crc kubenswrapper[4840]: I0930 13:58:46.535916 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:46 crc kubenswrapper[4840]: E0930 13:58:46.536315 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:47.036298963 +0000 UTC m=+155.665385386 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:46 crc kubenswrapper[4840]: I0930 13:58:46.618134 4840 patch_prober.go:28] interesting pod/router-default-5444994796-s9lx8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 13:58:46 crc kubenswrapper[4840]: [-]has-synced failed: reason withheld Sep 30 13:58:46 crc kubenswrapper[4840]: [+]process-running ok Sep 30 13:58:46 crc kubenswrapper[4840]: healthz check failed Sep 30 13:58:46 crc kubenswrapper[4840]: I0930 13:58:46.618204 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s9lx8" podUID="f0515176-8bf5-46a4-a905-7d308f64449e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 13:58:46 crc kubenswrapper[4840]: I0930 13:58:46.636789 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:46 crc kubenswrapper[4840]: E0930 13:58:46.636963 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:47.136933243 +0000 UTC m=+155.766019676 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:46 crc kubenswrapper[4840]: I0930 13:58:46.637175 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:46 crc kubenswrapper[4840]: E0930 13:58:46.637522 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:47.137512938 +0000 UTC m=+155.766599361 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:46 crc kubenswrapper[4840]: I0930 13:58:46.643088 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-56p76" event={"ID":"c4ecdaa2-b6f9-4682-b115-eb12a0060263","Type":"ContainerStarted","Data":"787c2143d0890af1f1710c6c655a42f9ba2d2e123c799bbba1a7a5cb2357d9e6"} Sep 30 13:58:46 crc kubenswrapper[4840]: I0930 13:58:46.739079 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:46 crc kubenswrapper[4840]: E0930 13:58:46.741075 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:47.241056884 +0000 UTC m=+155.870143307 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:46 crc kubenswrapper[4840]: I0930 13:58:46.841101 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:46 crc kubenswrapper[4840]: E0930 13:58:46.841811 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:47.341791706 +0000 UTC m=+155.970878139 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:46 crc kubenswrapper[4840]: I0930 13:58:46.942802 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:46 crc kubenswrapper[4840]: E0930 13:58:46.943180 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:47.443165455 +0000 UTC m=+156.072251878 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.044698 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.044791 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.044836 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.044874 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.044934 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:58:47 crc kubenswrapper[4840]: E0930 13:58:47.045337 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:47.545322955 +0000 UTC m=+156.174409378 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.047314 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.051676 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.054716 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.058990 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.128445 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.147243 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:47 crc kubenswrapper[4840]: E0930 13:58:47.147452 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:47.647428443 +0000 UTC m=+156.276514866 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.147585 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:47 crc kubenswrapper[4840]: E0930 13:58:47.147917 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:47.647906906 +0000 UTC m=+156.276993329 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.237628 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.248197 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:47 crc kubenswrapper[4840]: E0930 13:58:47.248628 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:47.748612357 +0000 UTC m=+156.377698780 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.251789 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.294062 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9s79z" Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.349325 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:47 crc kubenswrapper[4840]: E0930 13:58:47.349674 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:47.849662258 +0000 UTC m=+156.478748681 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.453120 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:47 crc kubenswrapper[4840]: E0930 13:58:47.453478 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:47.953463481 +0000 UTC m=+156.582549904 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.534540 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qxvn2"] Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.535735 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qxvn2" Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.542987 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.556780 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:47 crc kubenswrapper[4840]: E0930 13:58:47.557078 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:48.057066739 +0000 UTC m=+156.686153162 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.578452 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qxvn2"] Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.621815 4840 patch_prober.go:28] interesting pod/router-default-5444994796-s9lx8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 13:58:47 crc kubenswrapper[4840]: [-]has-synced failed: reason withheld Sep 30 13:58:47 crc kubenswrapper[4840]: [+]process-running ok Sep 30 13:58:47 crc kubenswrapper[4840]: healthz check failed Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.621871 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s9lx8" podUID="f0515176-8bf5-46a4-a905-7d308f64449e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.650684 4840 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-n2cvx container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.650752 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n2cvx" podUID="f5ac3e36-a48a-477a-b83e-6d9acaba6169" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.32:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.653694 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tc2fx"] Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.654593 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tc2fx" Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.657967 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.658169 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c2b4d42-9935-4e56-972f-cdd95d5a0e86-utilities\") pod \"community-operators-qxvn2\" (UID: \"3c2b4d42-9935-4e56-972f-cdd95d5a0e86\") " pod="openshift-marketplace/community-operators-qxvn2" Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.658207 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c2b4d42-9935-4e56-972f-cdd95d5a0e86-catalog-content\") pod \"community-operators-qxvn2\" (UID: \"3c2b4d42-9935-4e56-972f-cdd95d5a0e86\") " pod="openshift-marketplace/community-operators-qxvn2" Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.658237 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bt794\" (UniqueName: \"kubernetes.io/projected/3c2b4d42-9935-4e56-972f-cdd95d5a0e86-kube-api-access-bt794\") pod \"community-operators-qxvn2\" (UID: \"3c2b4d42-9935-4e56-972f-cdd95d5a0e86\") " pod="openshift-marketplace/community-operators-qxvn2" Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.658435 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Sep 30 13:58:47 crc kubenswrapper[4840]: E0930 13:58:47.658632 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:48.158618523 +0000 UTC m=+156.787704946 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.663302 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-56p76" event={"ID":"c4ecdaa2-b6f9-4682-b115-eb12a0060263","Type":"ContainerStarted","Data":"229a899f295580754362cbd8ba06fc86b5b6331641b7f49b5bb8181e4ad9e94d"} Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.761192 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.761226 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c2b4d42-9935-4e56-972f-cdd95d5a0e86-catalog-content\") pod \"community-operators-qxvn2\" (UID: \"3c2b4d42-9935-4e56-972f-cdd95d5a0e86\") " pod="openshift-marketplace/community-operators-qxvn2" Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.761276 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxk8d\" (UniqueName: \"kubernetes.io/projected/c97b5b98-9a09-4c3b-9bd0-5a5a8671c479-kube-api-access-nxk8d\") pod \"certified-operators-tc2fx\" (UID: \"c97b5b98-9a09-4c3b-9bd0-5a5a8671c479\") " pod="openshift-marketplace/certified-operators-tc2fx" Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.761292 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bt794\" (UniqueName: \"kubernetes.io/projected/3c2b4d42-9935-4e56-972f-cdd95d5a0e86-kube-api-access-bt794\") pod \"community-operators-qxvn2\" (UID: \"3c2b4d42-9935-4e56-972f-cdd95d5a0e86\") " pod="openshift-marketplace/community-operators-qxvn2" Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.761326 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c97b5b98-9a09-4c3b-9bd0-5a5a8671c479-catalog-content\") pod \"certified-operators-tc2fx\" (UID: \"c97b5b98-9a09-4c3b-9bd0-5a5a8671c479\") " pod="openshift-marketplace/certified-operators-tc2fx" Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.761362 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c97b5b98-9a09-4c3b-9bd0-5a5a8671c479-utilities\") pod \"certified-operators-tc2fx\" (UID: \"c97b5b98-9a09-4c3b-9bd0-5a5a8671c479\") " pod="openshift-marketplace/certified-operators-tc2fx" Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.761620 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c2b4d42-9935-4e56-972f-cdd95d5a0e86-utilities\") pod \"community-operators-qxvn2\" (UID: \"3c2b4d42-9935-4e56-972f-cdd95d5a0e86\") " pod="openshift-marketplace/community-operators-qxvn2" Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.762262 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c2b4d42-9935-4e56-972f-cdd95d5a0e86-utilities\") pod \"community-operators-qxvn2\" (UID: \"3c2b4d42-9935-4e56-972f-cdd95d5a0e86\") " pod="openshift-marketplace/community-operators-qxvn2" Sep 30 13:58:47 crc kubenswrapper[4840]: E0930 13:58:47.763426 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:48.263410502 +0000 UTC m=+156.892496925 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.763969 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c2b4d42-9935-4e56-972f-cdd95d5a0e86-catalog-content\") pod \"community-operators-qxvn2\" (UID: \"3c2b4d42-9935-4e56-972f-cdd95d5a0e86\") " pod="openshift-marketplace/community-operators-qxvn2" Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.791477 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tc2fx"] Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.814871 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bt794\" (UniqueName: \"kubernetes.io/projected/3c2b4d42-9935-4e56-972f-cdd95d5a0e86-kube-api-access-bt794\") pod \"community-operators-qxvn2\" (UID: \"3c2b4d42-9935-4e56-972f-cdd95d5a0e86\") " pod="openshift-marketplace/community-operators-qxvn2" Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.843367 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rdflm"] Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.844385 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rdflm" Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.866149 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.866288 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c97b5b98-9a09-4c3b-9bd0-5a5a8671c479-catalog-content\") pod \"certified-operators-tc2fx\" (UID: \"c97b5b98-9a09-4c3b-9bd0-5a5a8671c479\") " pod="openshift-marketplace/certified-operators-tc2fx" Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.866316 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c97b5b98-9a09-4c3b-9bd0-5a5a8671c479-utilities\") pod \"certified-operators-tc2fx\" (UID: \"c97b5b98-9a09-4c3b-9bd0-5a5a8671c479\") " pod="openshift-marketplace/certified-operators-tc2fx" Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.866394 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxk8d\" (UniqueName: \"kubernetes.io/projected/c97b5b98-9a09-4c3b-9bd0-5a5a8671c479-kube-api-access-nxk8d\") pod \"certified-operators-tc2fx\" (UID: \"c97b5b98-9a09-4c3b-9bd0-5a5a8671c479\") " pod="openshift-marketplace/certified-operators-tc2fx" Sep 30 13:58:47 crc kubenswrapper[4840]: E0930 13:58:47.866716 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:48.366701692 +0000 UTC m=+156.995788105 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.867014 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c97b5b98-9a09-4c3b-9bd0-5a5a8671c479-catalog-content\") pod \"certified-operators-tc2fx\" (UID: \"c97b5b98-9a09-4c3b-9bd0-5a5a8671c479\") " pod="openshift-marketplace/certified-operators-tc2fx" Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.867216 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c97b5b98-9a09-4c3b-9bd0-5a5a8671c479-utilities\") pod \"certified-operators-tc2fx\" (UID: \"c97b5b98-9a09-4c3b-9bd0-5a5a8671c479\") " pod="openshift-marketplace/certified-operators-tc2fx" Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.872263 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rdflm"] Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.872578 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qxvn2" Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.929542 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxk8d\" (UniqueName: \"kubernetes.io/projected/c97b5b98-9a09-4c3b-9bd0-5a5a8671c479-kube-api-access-nxk8d\") pod \"certified-operators-tc2fx\" (UID: \"c97b5b98-9a09-4c3b-9bd0-5a5a8671c479\") " pod="openshift-marketplace/certified-operators-tc2fx" Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.967221 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4-utilities\") pod \"community-operators-rdflm\" (UID: \"a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4\") " pod="openshift-marketplace/community-operators-rdflm" Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.967279 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4-catalog-content\") pod \"community-operators-rdflm\" (UID: \"a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4\") " pod="openshift-marketplace/community-operators-rdflm" Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.967326 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.967361 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fddm9\" (UniqueName: \"kubernetes.io/projected/a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4-kube-api-access-fddm9\") pod \"community-operators-rdflm\" (UID: \"a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4\") " pod="openshift-marketplace/community-operators-rdflm" Sep 30 13:58:47 crc kubenswrapper[4840]: E0930 13:58:47.968147 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:48.468134833 +0000 UTC m=+157.097221256 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:47 crc kubenswrapper[4840]: I0930 13:58:47.982207 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tc2fx" Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.048081 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dd72s"] Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.053637 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dd72s" Sep 30 13:58:48 crc kubenswrapper[4840]: W0930 13:58:48.054977 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-efa310e496f19da1e56825e4c7e014bdc5ea21bd7f120a6aaa1c03d554329707 WatchSource:0}: Error finding container efa310e496f19da1e56825e4c7e014bdc5ea21bd7f120a6aaa1c03d554329707: Status 404 returned error can't find the container with id efa310e496f19da1e56825e4c7e014bdc5ea21bd7f120a6aaa1c03d554329707 Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.068123 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.068403 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fddm9\" (UniqueName: \"kubernetes.io/projected/a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4-kube-api-access-fddm9\") pod \"community-operators-rdflm\" (UID: \"a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4\") " pod="openshift-marketplace/community-operators-rdflm" Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.068468 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4-utilities\") pod \"community-operators-rdflm\" (UID: \"a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4\") " pod="openshift-marketplace/community-operators-rdflm" Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.068502 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4-catalog-content\") pod \"community-operators-rdflm\" (UID: \"a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4\") " pod="openshift-marketplace/community-operators-rdflm" Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.069020 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4-catalog-content\") pod \"community-operators-rdflm\" (UID: \"a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4\") " pod="openshift-marketplace/community-operators-rdflm" Sep 30 13:58:48 crc kubenswrapper[4840]: E0930 13:58:48.069131 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:48.569079771 +0000 UTC m=+157.198166194 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.069693 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4-utilities\") pod \"community-operators-rdflm\" (UID: \"a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4\") " pod="openshift-marketplace/community-operators-rdflm" Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.105312 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dd72s"] Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.126105 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fddm9\" (UniqueName: \"kubernetes.io/projected/a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4-kube-api-access-fddm9\") pod \"community-operators-rdflm\" (UID: \"a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4\") " pod="openshift-marketplace/community-operators-rdflm" Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.171832 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcd2c899-747f-42ed-b1c4-cf6b70b5b230-catalog-content\") pod \"certified-operators-dd72s\" (UID: \"fcd2c899-747f-42ed-b1c4-cf6b70b5b230\") " pod="openshift-marketplace/certified-operators-dd72s" Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.172244 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.172276 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzr4v\" (UniqueName: \"kubernetes.io/projected/fcd2c899-747f-42ed-b1c4-cf6b70b5b230-kube-api-access-lzr4v\") pod \"certified-operators-dd72s\" (UID: \"fcd2c899-747f-42ed-b1c4-cf6b70b5b230\") " pod="openshift-marketplace/certified-operators-dd72s" Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.172307 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcd2c899-747f-42ed-b1c4-cf6b70b5b230-utilities\") pod \"certified-operators-dd72s\" (UID: \"fcd2c899-747f-42ed-b1c4-cf6b70b5b230\") " pod="openshift-marketplace/certified-operators-dd72s" Sep 30 13:58:48 crc kubenswrapper[4840]: E0930 13:58:48.172659 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:48.672646608 +0000 UTC m=+157.301733031 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.214888 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rdflm" Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.280609 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.280844 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzr4v\" (UniqueName: \"kubernetes.io/projected/fcd2c899-747f-42ed-b1c4-cf6b70b5b230-kube-api-access-lzr4v\") pod \"certified-operators-dd72s\" (UID: \"fcd2c899-747f-42ed-b1c4-cf6b70b5b230\") " pod="openshift-marketplace/certified-operators-dd72s" Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.280893 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcd2c899-747f-42ed-b1c4-cf6b70b5b230-utilities\") pod \"certified-operators-dd72s\" (UID: \"fcd2c899-747f-42ed-b1c4-cf6b70b5b230\") " pod="openshift-marketplace/certified-operators-dd72s" Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.280927 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcd2c899-747f-42ed-b1c4-cf6b70b5b230-catalog-content\") pod \"certified-operators-dd72s\" (UID: \"fcd2c899-747f-42ed-b1c4-cf6b70b5b230\") " pod="openshift-marketplace/certified-operators-dd72s" Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.281340 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcd2c899-747f-42ed-b1c4-cf6b70b5b230-catalog-content\") pod \"certified-operators-dd72s\" (UID: \"fcd2c899-747f-42ed-b1c4-cf6b70b5b230\") " pod="openshift-marketplace/certified-operators-dd72s" Sep 30 13:58:48 crc kubenswrapper[4840]: E0930 13:58:48.281403 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:48.781388311 +0000 UTC m=+157.410474734 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.281929 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcd2c899-747f-42ed-b1c4-cf6b70b5b230-utilities\") pod \"certified-operators-dd72s\" (UID: \"fcd2c899-747f-42ed-b1c4-cf6b70b5b230\") " pod="openshift-marketplace/certified-operators-dd72s" Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.364398 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzr4v\" (UniqueName: \"kubernetes.io/projected/fcd2c899-747f-42ed-b1c4-cf6b70b5b230-kube-api-access-lzr4v\") pod \"certified-operators-dd72s\" (UID: \"fcd2c899-747f-42ed-b1c4-cf6b70b5b230\") " pod="openshift-marketplace/certified-operators-dd72s" Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.387708 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:48 crc kubenswrapper[4840]: E0930 13:58:48.388339 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:48.888327368 +0000 UTC m=+157.517413791 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.454867 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dd72s" Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.489293 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:48 crc kubenswrapper[4840]: E0930 13:58:48.489388 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:48.989374028 +0000 UTC m=+157.618460451 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.489625 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:48 crc kubenswrapper[4840]: E0930 13:58:48.489943 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:48.989936103 +0000 UTC m=+157.619022526 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.564954 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qxvn2"] Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.593104 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:48 crc kubenswrapper[4840]: E0930 13:58:48.593498 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:49.093478499 +0000 UTC m=+157.722564922 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.618823 4840 patch_prober.go:28] interesting pod/router-default-5444994796-s9lx8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 13:58:48 crc kubenswrapper[4840]: [-]has-synced failed: reason withheld Sep 30 13:58:48 crc kubenswrapper[4840]: [+]process-running ok Sep 30 13:58:48 crc kubenswrapper[4840]: healthz check failed Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.619112 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s9lx8" podUID="f0515176-8bf5-46a4-a905-7d308f64449e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.665355 4840 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-n2cvx container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.665433 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n2cvx" podUID="f5ac3e36-a48a-477a-b83e-6d9acaba6169" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.32:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.697664 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:48 crc kubenswrapper[4840]: E0930 13:58:48.697980 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:49.197968671 +0000 UTC m=+157.827055094 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.732184 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"e54933899a75abaa8f2070fa7b7da3e630b04794e5ee97abc31add4d7c2d4a31"} Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.732226 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"e7203fc304acf6969662afce21d59846b650fae22f1c0c7497e145da5f5d14cc"} Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.732444 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.765702 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-56p76" event={"ID":"c4ecdaa2-b6f9-4682-b115-eb12a0060263","Type":"ContainerStarted","Data":"202d57c337b4104834800d504b760eee1a85ea206d735a864fff10d2a7e99816"} Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.773187 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tc2fx"] Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.773362 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qxvn2" event={"ID":"3c2b4d42-9935-4e56-972f-cdd95d5a0e86","Type":"ContainerStarted","Data":"a4b2eff50a8ea900a24e4a66c753298952e6b102215140619f8a8f8e6e7b49b9"} Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.777410 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"efa310e496f19da1e56825e4c7e014bdc5ea21bd7f120a6aaa1c03d554329707"} Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.778595 4840 generic.go:334] "Generic (PLEG): container finished" podID="850a2c18-882c-495f-a7e8-743e425a8b61" containerID="bc5d3b48314ba858d8d83c52b612834e9bda381052fd43502a08ffde2724327d" exitCode=0 Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.778636 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320665-rkbh5" event={"ID":"850a2c18-882c-495f-a7e8-743e425a8b61","Type":"ContainerDied","Data":"bc5d3b48314ba858d8d83c52b612834e9bda381052fd43502a08ffde2724327d"} Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.779526 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"825ff9837c3f5aa0c667b40d273b5f1229293d1f983bb0d27184a05eb8b1f06c"} Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.800603 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:48 crc kubenswrapper[4840]: E0930 13:58:48.801338 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:49.301312532 +0000 UTC m=+157.930398955 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:48 crc kubenswrapper[4840]: I0930 13:58:48.904330 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:48 crc kubenswrapper[4840]: E0930 13:58:48.904721 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:49.404707665 +0000 UTC m=+158.033794088 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.007479 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:49 crc kubenswrapper[4840]: E0930 13:58:49.007839 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:49.50782153 +0000 UTC m=+158.136907953 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.008041 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:49 crc kubenswrapper[4840]: E0930 13:58:49.008305 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:49.508298112 +0000 UTC m=+158.137384535 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.108943 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:49 crc kubenswrapper[4840]: E0930 13:58:49.109141 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:49.609115997 +0000 UTC m=+158.238202420 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.109217 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:49 crc kubenswrapper[4840]: E0930 13:58:49.109760 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:49.609742293 +0000 UTC m=+158.238828716 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.196141 4840 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.210072 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:49 crc kubenswrapper[4840]: E0930 13:58:49.210632 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 13:58:49.710613579 +0000 UTC m=+158.339700002 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.241162 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dd72s"] Sep 30 13:58:49 crc kubenswrapper[4840]: W0930 13:58:49.260734 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfcd2c899_747f_42ed_b1c4_cf6b70b5b230.slice/crio-035c917febad21cbcf2d1a650107529376c542b8f40b7dc2e5c6b10b1c62d8c2 WatchSource:0}: Error finding container 035c917febad21cbcf2d1a650107529376c542b8f40b7dc2e5c6b10b1c62d8c2: Status 404 returned error can't find the container with id 035c917febad21cbcf2d1a650107529376c542b8f40b7dc2e5c6b10b1c62d8c2 Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.311263 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:49 crc kubenswrapper[4840]: E0930 13:58:49.311604 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 13:58:49.811592798 +0000 UTC m=+158.440679221 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lzwls" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.334169 4840 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-09-30T13:58:49.196178778Z","Handler":null,"Name":""} Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.340107 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rdflm"] Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.344760 4840 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.344798 4840 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Sep 30 13:58:49 crc kubenswrapper[4840]: W0930 13:58:49.348457 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda2ec4007_7bc3_43dd_bbe5_b172bb7b30e4.slice/crio-e8c21dcafd69f1fd2c21f53ee7d5d0f95b6ff7fff97a4739f6154b7747e2bea7 WatchSource:0}: Error finding container e8c21dcafd69f1fd2c21f53ee7d5d0f95b6ff7fff97a4739f6154b7747e2bea7: Status 404 returned error can't find the container with id e8c21dcafd69f1fd2c21f53ee7d5d0f95b6ff7fff97a4739f6154b7747e2bea7 Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.412646 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.422300 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.432707 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-z895r"] Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.433704 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z895r" Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.437203 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.447221 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z895r"] Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.513840 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d587bdc-9a81-4e7c-9df2-d638d43f5e0d-utilities\") pod \"redhat-marketplace-z895r\" (UID: \"5d587bdc-9a81-4e7c-9df2-d638d43f5e0d\") " pod="openshift-marketplace/redhat-marketplace-z895r" Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.513887 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d587bdc-9a81-4e7c-9df2-d638d43f5e0d-catalog-content\") pod \"redhat-marketplace-z895r\" (UID: \"5d587bdc-9a81-4e7c-9df2-d638d43f5e0d\") " pod="openshift-marketplace/redhat-marketplace-z895r" Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.513938 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6t49m\" (UniqueName: \"kubernetes.io/projected/5d587bdc-9a81-4e7c-9df2-d638d43f5e0d-kube-api-access-6t49m\") pod \"redhat-marketplace-z895r\" (UID: \"5d587bdc-9a81-4e7c-9df2-d638d43f5e0d\") " pod="openshift-marketplace/redhat-marketplace-z895r" Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.513965 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.516746 4840 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.516787 4840 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.544016 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lzwls\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.614578 4840 patch_prober.go:28] interesting pod/router-default-5444994796-s9lx8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 13:58:49 crc kubenswrapper[4840]: [-]has-synced failed: reason withheld Sep 30 13:58:49 crc kubenswrapper[4840]: [+]process-running ok Sep 30 13:58:49 crc kubenswrapper[4840]: healthz check failed Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.614647 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s9lx8" podUID="f0515176-8bf5-46a4-a905-7d308f64449e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.614672 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d587bdc-9a81-4e7c-9df2-d638d43f5e0d-utilities\") pod \"redhat-marketplace-z895r\" (UID: \"5d587bdc-9a81-4e7c-9df2-d638d43f5e0d\") " pod="openshift-marketplace/redhat-marketplace-z895r" Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.614721 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d587bdc-9a81-4e7c-9df2-d638d43f5e0d-catalog-content\") pod \"redhat-marketplace-z895r\" (UID: \"5d587bdc-9a81-4e7c-9df2-d638d43f5e0d\") " pod="openshift-marketplace/redhat-marketplace-z895r" Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.614791 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6t49m\" (UniqueName: \"kubernetes.io/projected/5d587bdc-9a81-4e7c-9df2-d638d43f5e0d-kube-api-access-6t49m\") pod \"redhat-marketplace-z895r\" (UID: \"5d587bdc-9a81-4e7c-9df2-d638d43f5e0d\") " pod="openshift-marketplace/redhat-marketplace-z895r" Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.615152 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d587bdc-9a81-4e7c-9df2-d638d43f5e0d-utilities\") pod \"redhat-marketplace-z895r\" (UID: \"5d587bdc-9a81-4e7c-9df2-d638d43f5e0d\") " pod="openshift-marketplace/redhat-marketplace-z895r" Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.615220 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d587bdc-9a81-4e7c-9df2-d638d43f5e0d-catalog-content\") pod \"redhat-marketplace-z895r\" (UID: \"5d587bdc-9a81-4e7c-9df2-d638d43f5e0d\") " pod="openshift-marketplace/redhat-marketplace-z895r" Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.636666 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6t49m\" (UniqueName: \"kubernetes.io/projected/5d587bdc-9a81-4e7c-9df2-d638d43f5e0d-kube-api-access-6t49m\") pod \"redhat-marketplace-z895r\" (UID: \"5d587bdc-9a81-4e7c-9df2-d638d43f5e0d\") " pod="openshift-marketplace/redhat-marketplace-z895r" Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.691682 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.774070 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z895r" Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.789454 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"0263fa4d10f532d3f5bb2130b9c2b6fa7a4e38a7803ea6da8dcda4e3b9231338"} Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.798999 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"9a7de8569733439f293c56a252eb3aeb317a25557ab978dc13c5a578d129d9e4"} Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.808936 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-56p76" event={"ID":"c4ecdaa2-b6f9-4682-b115-eb12a0060263","Type":"ContainerStarted","Data":"3ab1cd42bdd5f6d364df017a07d99b69a47e86d00ccf8b18c65eb5bafecfbc93"} Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.811293 4840 generic.go:334] "Generic (PLEG): container finished" podID="3c2b4d42-9935-4e56-972f-cdd95d5a0e86" containerID="d99911001fbf445dedee82d6b2eeabfcc62f08e815da035535039377331df036" exitCode=0 Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.811368 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qxvn2" event={"ID":"3c2b4d42-9935-4e56-972f-cdd95d5a0e86","Type":"ContainerDied","Data":"d99911001fbf445dedee82d6b2eeabfcc62f08e815da035535039377331df036"} Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.813976 4840 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.816277 4840 generic.go:334] "Generic (PLEG): container finished" podID="c97b5b98-9a09-4c3b-9bd0-5a5a8671c479" containerID="0ed1236664362f6f4988de894b3dbaf9ce29ab242e2a03f872b438a16b279af7" exitCode=0 Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.816377 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tc2fx" event={"ID":"c97b5b98-9a09-4c3b-9bd0-5a5a8671c479","Type":"ContainerDied","Data":"0ed1236664362f6f4988de894b3dbaf9ce29ab242e2a03f872b438a16b279af7"} Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.816449 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tc2fx" event={"ID":"c97b5b98-9a09-4c3b-9bd0-5a5a8671c479","Type":"ContainerStarted","Data":"71b299bc9ebe1e1e182a3e5e3f8a058663b44eeb9aac7e2442d4eb684c0025b2"} Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.823230 4840 generic.go:334] "Generic (PLEG): container finished" podID="a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4" containerID="1edd4f8f22d64d076189d4949a8d935a45780a20d367f23ac4e9f35ef6d668c2" exitCode=0 Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.823905 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rdflm" event={"ID":"a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4","Type":"ContainerDied","Data":"1edd4f8f22d64d076189d4949a8d935a45780a20d367f23ac4e9f35ef6d668c2"} Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.823929 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rdflm" event={"ID":"a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4","Type":"ContainerStarted","Data":"e8c21dcafd69f1fd2c21f53ee7d5d0f95b6ff7fff97a4739f6154b7747e2bea7"} Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.837206 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xn8ts"] Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.839523 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xn8ts" Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.843109 4840 generic.go:334] "Generic (PLEG): container finished" podID="fcd2c899-747f-42ed-b1c4-cf6b70b5b230" containerID="7ae770494d9a9fac05d23e9accc3db4865620fe0c704e646230a1e69b61473d8" exitCode=0 Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.843189 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dd72s" event={"ID":"fcd2c899-747f-42ed-b1c4-cf6b70b5b230","Type":"ContainerDied","Data":"7ae770494d9a9fac05d23e9accc3db4865620fe0c704e646230a1e69b61473d8"} Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.843239 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dd72s" event={"ID":"fcd2c899-747f-42ed-b1c4-cf6b70b5b230","Type":"ContainerStarted","Data":"035c917febad21cbcf2d1a650107529376c542b8f40b7dc2e5c6b10b1c62d8c2"} Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.853314 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xn8ts"] Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.885366 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-56p76" podStartSLOduration=11.885333681 podStartE2EDuration="11.885333681s" podCreationTimestamp="2025-09-30 13:58:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:49.882901756 +0000 UTC m=+158.511988189" watchObservedRunningTime="2025-09-30 13:58:49.885333681 +0000 UTC m=+158.514420104" Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.928995 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/796c97de-5fd5-4ffa-886b-3ffa2ead1da3-utilities\") pod \"redhat-marketplace-xn8ts\" (UID: \"796c97de-5fd5-4ffa-886b-3ffa2ead1da3\") " pod="openshift-marketplace/redhat-marketplace-xn8ts" Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.929143 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/796c97de-5fd5-4ffa-886b-3ffa2ead1da3-catalog-content\") pod \"redhat-marketplace-xn8ts\" (UID: \"796c97de-5fd5-4ffa-886b-3ffa2ead1da3\") " pod="openshift-marketplace/redhat-marketplace-xn8ts" Sep 30 13:58:49 crc kubenswrapper[4840]: I0930 13:58:49.929194 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9ljk\" (UniqueName: \"kubernetes.io/projected/796c97de-5fd5-4ffa-886b-3ffa2ead1da3-kube-api-access-p9ljk\") pod \"redhat-marketplace-xn8ts\" (UID: \"796c97de-5fd5-4ffa-886b-3ffa2ead1da3\") " pod="openshift-marketplace/redhat-marketplace-xn8ts" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.030824 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/796c97de-5fd5-4ffa-886b-3ffa2ead1da3-utilities\") pod \"redhat-marketplace-xn8ts\" (UID: \"796c97de-5fd5-4ffa-886b-3ffa2ead1da3\") " pod="openshift-marketplace/redhat-marketplace-xn8ts" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.030889 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/796c97de-5fd5-4ffa-886b-3ffa2ead1da3-catalog-content\") pod \"redhat-marketplace-xn8ts\" (UID: \"796c97de-5fd5-4ffa-886b-3ffa2ead1da3\") " pod="openshift-marketplace/redhat-marketplace-xn8ts" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.030918 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9ljk\" (UniqueName: \"kubernetes.io/projected/796c97de-5fd5-4ffa-886b-3ffa2ead1da3-kube-api-access-p9ljk\") pod \"redhat-marketplace-xn8ts\" (UID: \"796c97de-5fd5-4ffa-886b-3ffa2ead1da3\") " pod="openshift-marketplace/redhat-marketplace-xn8ts" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.031360 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/796c97de-5fd5-4ffa-886b-3ffa2ead1da3-utilities\") pod \"redhat-marketplace-xn8ts\" (UID: \"796c97de-5fd5-4ffa-886b-3ffa2ead1da3\") " pod="openshift-marketplace/redhat-marketplace-xn8ts" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.031442 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/796c97de-5fd5-4ffa-886b-3ffa2ead1da3-catalog-content\") pod \"redhat-marketplace-xn8ts\" (UID: \"796c97de-5fd5-4ffa-886b-3ffa2ead1da3\") " pod="openshift-marketplace/redhat-marketplace-xn8ts" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.048373 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9ljk\" (UniqueName: \"kubernetes.io/projected/796c97de-5fd5-4ffa-886b-3ffa2ead1da3-kube-api-access-p9ljk\") pod \"redhat-marketplace-xn8ts\" (UID: \"796c97de-5fd5-4ffa-886b-3ffa2ead1da3\") " pod="openshift-marketplace/redhat-marketplace-xn8ts" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.069216 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z895r"] Sep 30 13:58:50 crc kubenswrapper[4840]: W0930 13:58:50.080591 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5d587bdc_9a81_4e7c_9df2_d638d43f5e0d.slice/crio-27587a514f130738ead13f9bc8547449515e24e6965c0f9263e18cdfe4b1f107 WatchSource:0}: Error finding container 27587a514f130738ead13f9bc8547449515e24e6965c0f9263e18cdfe4b1f107: Status 404 returned error can't find the container with id 27587a514f130738ead13f9bc8547449515e24e6965c0f9263e18cdfe4b1f107 Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.134090 4840 patch_prober.go:28] interesting pod/downloads-7954f5f757-kdwqh container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.134149 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-kdwqh" podUID="dde0d17e-6bd0-4acb-b8b8-a8b0bc988a6e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.134451 4840 patch_prober.go:28] interesting pod/downloads-7954f5f757-kdwqh container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.134621 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-kdwqh" podUID="dde0d17e-6bd0-4acb-b8b8-a8b0bc988a6e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.145707 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.146273 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-lzwls"] Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.172011 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xn8ts" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.221883 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320665-rkbh5" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.336155 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/850a2c18-882c-495f-a7e8-743e425a8b61-secret-volume\") pod \"850a2c18-882c-495f-a7e8-743e425a8b61\" (UID: \"850a2c18-882c-495f-a7e8-743e425a8b61\") " Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.336232 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gkrg5\" (UniqueName: \"kubernetes.io/projected/850a2c18-882c-495f-a7e8-743e425a8b61-kube-api-access-gkrg5\") pod \"850a2c18-882c-495f-a7e8-743e425a8b61\" (UID: \"850a2c18-882c-495f-a7e8-743e425a8b61\") " Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.336467 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/850a2c18-882c-495f-a7e8-743e425a8b61-config-volume\") pod \"850a2c18-882c-495f-a7e8-743e425a8b61\" (UID: \"850a2c18-882c-495f-a7e8-743e425a8b61\") " Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.337631 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/850a2c18-882c-495f-a7e8-743e425a8b61-config-volume" (OuterVolumeSpecName: "config-volume") pod "850a2c18-882c-495f-a7e8-743e425a8b61" (UID: "850a2c18-882c-495f-a7e8-743e425a8b61"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.342648 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-wdjfz" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.343116 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/850a2c18-882c-495f-a7e8-743e425a8b61-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "850a2c18-882c-495f-a7e8-743e425a8b61" (UID: "850a2c18-882c-495f-a7e8-743e425a8b61"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.345357 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-qjxnf" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.345640 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-qjxnf" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.346431 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/850a2c18-882c-495f-a7e8-743e425a8b61-kube-api-access-gkrg5" (OuterVolumeSpecName: "kube-api-access-gkrg5") pod "850a2c18-882c-495f-a7e8-743e425a8b61" (UID: "850a2c18-882c-495f-a7e8-743e425a8b61"). InnerVolumeSpecName "kube-api-access-gkrg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.356799 4840 patch_prober.go:28] interesting pod/console-f9d7485db-qjxnf container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.6:8443/health\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.356847 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-qjxnf" podUID="9dd57b46-77fc-4b8a-9fc1-a50084165914" containerName="console" probeResult="failure" output="Get \"https://10.217.0.6:8443/health\": dial tcp 10.217.0.6:8443: connect: connection refused" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.384938 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xn8ts"] Sep 30 13:58:50 crc kubenswrapper[4840]: W0930 13:58:50.423319 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod796c97de_5fd5_4ffa_886b_3ffa2ead1da3.slice/crio-2ae45e4e3399b314482af5674bcf5bf6fe6499c8c6499db52fad547803d6bbca WatchSource:0}: Error finding container 2ae45e4e3399b314482af5674bcf5bf6fe6499c8c6499db52fad547803d6bbca: Status 404 returned error can't find the container with id 2ae45e4e3399b314482af5674bcf5bf6fe6499c8c6499db52fad547803d6bbca Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.437988 4840 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/850a2c18-882c-495f-a7e8-743e425a8b61-config-volume\") on node \"crc\" DevicePath \"\"" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.438021 4840 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/850a2c18-882c-495f-a7e8-743e425a8b61-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.438030 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gkrg5\" (UniqueName: \"kubernetes.io/projected/850a2c18-882c-495f-a7e8-743e425a8b61-kube-api-access-gkrg5\") on node \"crc\" DevicePath \"\"" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.570750 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.570806 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.578132 4840 patch_prober.go:28] interesting pod/apiserver-76f77b778f-lk4jg container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Sep 30 13:58:50 crc kubenswrapper[4840]: [+]log ok Sep 30 13:58:50 crc kubenswrapper[4840]: [+]etcd ok Sep 30 13:58:50 crc kubenswrapper[4840]: [+]poststarthook/start-apiserver-admission-initializer ok Sep 30 13:58:50 crc kubenswrapper[4840]: [+]poststarthook/generic-apiserver-start-informers ok Sep 30 13:58:50 crc kubenswrapper[4840]: [+]poststarthook/max-in-flight-filter ok Sep 30 13:58:50 crc kubenswrapper[4840]: [+]poststarthook/storage-object-count-tracker-hook ok Sep 30 13:58:50 crc kubenswrapper[4840]: [+]poststarthook/image.openshift.io-apiserver-caches ok Sep 30 13:58:50 crc kubenswrapper[4840]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Sep 30 13:58:50 crc kubenswrapper[4840]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Sep 30 13:58:50 crc kubenswrapper[4840]: [+]poststarthook/project.openshift.io-projectcache ok Sep 30 13:58:50 crc kubenswrapper[4840]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Sep 30 13:58:50 crc kubenswrapper[4840]: [+]poststarthook/openshift.io-startinformers ok Sep 30 13:58:50 crc kubenswrapper[4840]: [+]poststarthook/openshift.io-restmapperupdater ok Sep 30 13:58:50 crc kubenswrapper[4840]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Sep 30 13:58:50 crc kubenswrapper[4840]: livez check failed Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.578198 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" podUID="8ec02940-3b96-41aa-8624-8338ee544fe1" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.606326 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bf4gt" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.612463 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-s9lx8" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.613578 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bf4gt" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.616414 4840 patch_prober.go:28] interesting pod/router-default-5444994796-s9lx8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 13:58:50 crc kubenswrapper[4840]: [-]has-synced failed: reason withheld Sep 30 13:58:50 crc kubenswrapper[4840]: [+]process-running ok Sep 30 13:58:50 crc kubenswrapper[4840]: healthz check failed Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.616476 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s9lx8" podUID="f0515176-8bf5-46a4-a905-7d308f64449e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.841838 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5vk9z"] Sep 30 13:58:50 crc kubenswrapper[4840]: E0930 13:58:50.842080 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="850a2c18-882c-495f-a7e8-743e425a8b61" containerName="collect-profiles" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.842095 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="850a2c18-882c-495f-a7e8-743e425a8b61" containerName="collect-profiles" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.842215 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="850a2c18-882c-495f-a7e8-743e425a8b61" containerName="collect-profiles" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.843114 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5vk9z" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.845243 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.853781 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" event={"ID":"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067","Type":"ContainerStarted","Data":"ed46ee312f0c4b2993e82af345239da7c8960d19c03f6c04eb1b8ff14543fb43"} Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.853833 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" event={"ID":"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067","Type":"ContainerStarted","Data":"551e02e90f60f1e91c25acaf9d3e8f7b4aa8415f17aabd5be2e1641caabb54ca"} Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.854077 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.860852 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5vk9z"] Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.864697 4840 generic.go:334] "Generic (PLEG): container finished" podID="796c97de-5fd5-4ffa-886b-3ffa2ead1da3" containerID="f4c3cb6ee1ced67ef507f0dd198b81d6dee92f4c8c43786c75d9cd43e7723236" exitCode=0 Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.864823 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xn8ts" event={"ID":"796c97de-5fd5-4ffa-886b-3ffa2ead1da3","Type":"ContainerDied","Data":"f4c3cb6ee1ced67ef507f0dd198b81d6dee92f4c8c43786c75d9cd43e7723236"} Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.864860 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xn8ts" event={"ID":"796c97de-5fd5-4ffa-886b-3ffa2ead1da3","Type":"ContainerStarted","Data":"2ae45e4e3399b314482af5674bcf5bf6fe6499c8c6499db52fad547803d6bbca"} Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.867964 4840 generic.go:334] "Generic (PLEG): container finished" podID="5d587bdc-9a81-4e7c-9df2-d638d43f5e0d" containerID="b3db70d2b71ecaa2fbbdb1ffe2a875f5647437b38eae474dd5ffb08cc1b54a6d" exitCode=0 Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.868276 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z895r" event={"ID":"5d587bdc-9a81-4e7c-9df2-d638d43f5e0d","Type":"ContainerDied","Data":"b3db70d2b71ecaa2fbbdb1ffe2a875f5647437b38eae474dd5ffb08cc1b54a6d"} Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.868309 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z895r" event={"ID":"5d587bdc-9a81-4e7c-9df2-d638d43f5e0d","Type":"ContainerStarted","Data":"27587a514f130738ead13f9bc8547449515e24e6965c0f9263e18cdfe4b1f107"} Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.872243 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320665-rkbh5" event={"ID":"850a2c18-882c-495f-a7e8-743e425a8b61","Type":"ContainerDied","Data":"dc3ae9ae3faf80ae6dcd38bad5d98f623859292dcaa35b5b4151a7518ed73871"} Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.872282 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc3ae9ae3faf80ae6dcd38bad5d98f623859292dcaa35b5b4151a7518ed73871" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.872832 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320665-rkbh5" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.887405 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" podStartSLOduration=136.887385042 podStartE2EDuration="2m16.887385042s" podCreationTimestamp="2025-09-30 13:56:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:50.884978078 +0000 UTC m=+159.514064521" watchObservedRunningTime="2025-09-30 13:58:50.887385042 +0000 UTC m=+159.516471465" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.911085 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.912123 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.915292 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.915303 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.924429 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.945306 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f1c8962-d085-430f-9654-9aa798bc399c-utilities\") pod \"redhat-operators-5vk9z\" (UID: \"3f1c8962-d085-430f-9654-9aa798bc399c\") " pod="openshift-marketplace/redhat-operators-5vk9z" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.945386 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f1c8962-d085-430f-9654-9aa798bc399c-catalog-content\") pod \"redhat-operators-5vk9z\" (UID: \"3f1c8962-d085-430f-9654-9aa798bc399c\") " pod="openshift-marketplace/redhat-operators-5vk9z" Sep 30 13:58:50 crc kubenswrapper[4840]: I0930 13:58:50.945493 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x64f5\" (UniqueName: \"kubernetes.io/projected/3f1c8962-d085-430f-9654-9aa798bc399c-kube-api-access-x64f5\") pod \"redhat-operators-5vk9z\" (UID: \"3f1c8962-d085-430f-9654-9aa798bc399c\") " pod="openshift-marketplace/redhat-operators-5vk9z" Sep 30 13:58:51 crc kubenswrapper[4840]: I0930 13:58:51.043411 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-snznw" Sep 30 13:58:51 crc kubenswrapper[4840]: I0930 13:58:51.046303 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f1c8962-d085-430f-9654-9aa798bc399c-utilities\") pod \"redhat-operators-5vk9z\" (UID: \"3f1c8962-d085-430f-9654-9aa798bc399c\") " pod="openshift-marketplace/redhat-operators-5vk9z" Sep 30 13:58:51 crc kubenswrapper[4840]: I0930 13:58:51.046346 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f1c8962-d085-430f-9654-9aa798bc399c-catalog-content\") pod \"redhat-operators-5vk9z\" (UID: \"3f1c8962-d085-430f-9654-9aa798bc399c\") " pod="openshift-marketplace/redhat-operators-5vk9z" Sep 30 13:58:51 crc kubenswrapper[4840]: I0930 13:58:51.046402 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x64f5\" (UniqueName: \"kubernetes.io/projected/3f1c8962-d085-430f-9654-9aa798bc399c-kube-api-access-x64f5\") pod \"redhat-operators-5vk9z\" (UID: \"3f1c8962-d085-430f-9654-9aa798bc399c\") " pod="openshift-marketplace/redhat-operators-5vk9z" Sep 30 13:58:51 crc kubenswrapper[4840]: I0930 13:58:51.046455 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3c5d8d7f-1aa0-4ae5-98ca-b41d8c9946c1-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3c5d8d7f-1aa0-4ae5-98ca-b41d8c9946c1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 13:58:51 crc kubenswrapper[4840]: I0930 13:58:51.046485 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3c5d8d7f-1aa0-4ae5-98ca-b41d8c9946c1-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3c5d8d7f-1aa0-4ae5-98ca-b41d8c9946c1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 13:58:51 crc kubenswrapper[4840]: I0930 13:58:51.046997 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f1c8962-d085-430f-9654-9aa798bc399c-utilities\") pod \"redhat-operators-5vk9z\" (UID: \"3f1c8962-d085-430f-9654-9aa798bc399c\") " pod="openshift-marketplace/redhat-operators-5vk9z" Sep 30 13:58:51 crc kubenswrapper[4840]: I0930 13:58:51.047322 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f1c8962-d085-430f-9654-9aa798bc399c-catalog-content\") pod \"redhat-operators-5vk9z\" (UID: \"3f1c8962-d085-430f-9654-9aa798bc399c\") " pod="openshift-marketplace/redhat-operators-5vk9z" Sep 30 13:58:51 crc kubenswrapper[4840]: I0930 13:58:51.075427 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x64f5\" (UniqueName: \"kubernetes.io/projected/3f1c8962-d085-430f-9654-9aa798bc399c-kube-api-access-x64f5\") pod \"redhat-operators-5vk9z\" (UID: \"3f1c8962-d085-430f-9654-9aa798bc399c\") " pod="openshift-marketplace/redhat-operators-5vk9z" Sep 30 13:58:51 crc kubenswrapper[4840]: I0930 13:58:51.148255 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3c5d8d7f-1aa0-4ae5-98ca-b41d8c9946c1-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3c5d8d7f-1aa0-4ae5-98ca-b41d8c9946c1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 13:58:51 crc kubenswrapper[4840]: I0930 13:58:51.148322 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3c5d8d7f-1aa0-4ae5-98ca-b41d8c9946c1-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3c5d8d7f-1aa0-4ae5-98ca-b41d8c9946c1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 13:58:51 crc kubenswrapper[4840]: I0930 13:58:51.148726 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3c5d8d7f-1aa0-4ae5-98ca-b41d8c9946c1-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3c5d8d7f-1aa0-4ae5-98ca-b41d8c9946c1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 13:58:51 crc kubenswrapper[4840]: I0930 13:58:51.166175 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5vk9z" Sep 30 13:58:51 crc kubenswrapper[4840]: I0930 13:58:51.171924 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3c5d8d7f-1aa0-4ae5-98ca-b41d8c9946c1-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3c5d8d7f-1aa0-4ae5-98ca-b41d8c9946c1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 13:58:51 crc kubenswrapper[4840]: I0930 13:58:51.232766 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mfbwh"] Sep 30 13:58:51 crc kubenswrapper[4840]: I0930 13:58:51.234160 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mfbwh" Sep 30 13:58:51 crc kubenswrapper[4840]: I0930 13:58:51.244089 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mfbwh"] Sep 30 13:58:51 crc kubenswrapper[4840]: I0930 13:58:51.246346 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 13:58:51 crc kubenswrapper[4840]: I0930 13:58:51.324107 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n2cvx" Sep 30 13:58:51 crc kubenswrapper[4840]: I0930 13:58:51.355637 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrsbc\" (UniqueName: \"kubernetes.io/projected/96321113-bb05-4a57-b640-d706993c4520-kube-api-access-vrsbc\") pod \"redhat-operators-mfbwh\" (UID: \"96321113-bb05-4a57-b640-d706993c4520\") " pod="openshift-marketplace/redhat-operators-mfbwh" Sep 30 13:58:51 crc kubenswrapper[4840]: I0930 13:58:51.355738 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96321113-bb05-4a57-b640-d706993c4520-catalog-content\") pod \"redhat-operators-mfbwh\" (UID: \"96321113-bb05-4a57-b640-d706993c4520\") " pod="openshift-marketplace/redhat-operators-mfbwh" Sep 30 13:58:51 crc kubenswrapper[4840]: I0930 13:58:51.355765 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96321113-bb05-4a57-b640-d706993c4520-utilities\") pod \"redhat-operators-mfbwh\" (UID: \"96321113-bb05-4a57-b640-d706993c4520\") " pod="openshift-marketplace/redhat-operators-mfbwh" Sep 30 13:58:51 crc kubenswrapper[4840]: I0930 13:58:51.457342 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96321113-bb05-4a57-b640-d706993c4520-utilities\") pod \"redhat-operators-mfbwh\" (UID: \"96321113-bb05-4a57-b640-d706993c4520\") " pod="openshift-marketplace/redhat-operators-mfbwh" Sep 30 13:58:51 crc kubenswrapper[4840]: I0930 13:58:51.457410 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrsbc\" (UniqueName: \"kubernetes.io/projected/96321113-bb05-4a57-b640-d706993c4520-kube-api-access-vrsbc\") pod \"redhat-operators-mfbwh\" (UID: \"96321113-bb05-4a57-b640-d706993c4520\") " pod="openshift-marketplace/redhat-operators-mfbwh" Sep 30 13:58:51 crc kubenswrapper[4840]: I0930 13:58:51.457498 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96321113-bb05-4a57-b640-d706993c4520-catalog-content\") pod \"redhat-operators-mfbwh\" (UID: \"96321113-bb05-4a57-b640-d706993c4520\") " pod="openshift-marketplace/redhat-operators-mfbwh" Sep 30 13:58:51 crc kubenswrapper[4840]: I0930 13:58:51.458427 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96321113-bb05-4a57-b640-d706993c4520-catalog-content\") pod \"redhat-operators-mfbwh\" (UID: \"96321113-bb05-4a57-b640-d706993c4520\") " pod="openshift-marketplace/redhat-operators-mfbwh" Sep 30 13:58:51 crc kubenswrapper[4840]: I0930 13:58:51.458721 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96321113-bb05-4a57-b640-d706993c4520-utilities\") pod \"redhat-operators-mfbwh\" (UID: \"96321113-bb05-4a57-b640-d706993c4520\") " pod="openshift-marketplace/redhat-operators-mfbwh" Sep 30 13:58:51 crc kubenswrapper[4840]: I0930 13:58:51.480177 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrsbc\" (UniqueName: \"kubernetes.io/projected/96321113-bb05-4a57-b640-d706993c4520-kube-api-access-vrsbc\") pod \"redhat-operators-mfbwh\" (UID: \"96321113-bb05-4a57-b640-d706993c4520\") " pod="openshift-marketplace/redhat-operators-mfbwh" Sep 30 13:58:51 crc kubenswrapper[4840]: I0930 13:58:51.559774 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mfbwh" Sep 30 13:58:51 crc kubenswrapper[4840]: I0930 13:58:51.579840 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Sep 30 13:58:51 crc kubenswrapper[4840]: I0930 13:58:51.619211 4840 patch_prober.go:28] interesting pod/router-default-5444994796-s9lx8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 13:58:51 crc kubenswrapper[4840]: [-]has-synced failed: reason withheld Sep 30 13:58:51 crc kubenswrapper[4840]: [+]process-running ok Sep 30 13:58:51 crc kubenswrapper[4840]: healthz check failed Sep 30 13:58:51 crc kubenswrapper[4840]: I0930 13:58:51.619287 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s9lx8" podUID="f0515176-8bf5-46a4-a905-7d308f64449e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 13:58:51 crc kubenswrapper[4840]: I0930 13:58:51.631121 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-czbs2" Sep 30 13:58:51 crc kubenswrapper[4840]: I0930 13:58:51.651156 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5vk9z"] Sep 30 13:58:51 crc kubenswrapper[4840]: I0930 13:58:51.730440 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s72m9" Sep 30 13:58:51 crc kubenswrapper[4840]: I0930 13:58:51.871752 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 13:58:51 crc kubenswrapper[4840]: I0930 13:58:51.871804 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 13:58:51 crc kubenswrapper[4840]: I0930 13:58:51.883087 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3c5d8d7f-1aa0-4ae5-98ca-b41d8c9946c1","Type":"ContainerStarted","Data":"974c2ce5628eeffaf2764c1ef00ba0f126a42bdadd6130a88116275531de7a2f"} Sep 30 13:58:51 crc kubenswrapper[4840]: I0930 13:58:51.896910 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5vk9z" event={"ID":"3f1c8962-d085-430f-9654-9aa798bc399c","Type":"ContainerStarted","Data":"14c02f20e6ca97546431aa1e107c57bb3e48f643e7cac9f56ae269bd642a077f"} Sep 30 13:58:52 crc kubenswrapper[4840]: I0930 13:58:52.167144 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mfbwh"] Sep 30 13:58:52 crc kubenswrapper[4840]: I0930 13:58:52.615025 4840 patch_prober.go:28] interesting pod/router-default-5444994796-s9lx8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 13:58:52 crc kubenswrapper[4840]: [-]has-synced failed: reason withheld Sep 30 13:58:52 crc kubenswrapper[4840]: [+]process-running ok Sep 30 13:58:52 crc kubenswrapper[4840]: healthz check failed Sep 30 13:58:52 crc kubenswrapper[4840]: I0930 13:58:52.615086 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s9lx8" podUID="f0515176-8bf5-46a4-a905-7d308f64449e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 13:58:52 crc kubenswrapper[4840]: I0930 13:58:52.806639 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Sep 30 13:58:52 crc kubenswrapper[4840]: I0930 13:58:52.808202 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 13:58:52 crc kubenswrapper[4840]: I0930 13:58:52.813009 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Sep 30 13:58:52 crc kubenswrapper[4840]: I0930 13:58:52.813516 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Sep 30 13:58:52 crc kubenswrapper[4840]: I0930 13:58:52.814206 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Sep 30 13:58:52 crc kubenswrapper[4840]: I0930 13:58:52.891452 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6222d35a-eca2-4333-b0e4-0b45409c1645-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"6222d35a-eca2-4333-b0e4-0b45409c1645\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 13:58:52 crc kubenswrapper[4840]: I0930 13:58:52.891544 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6222d35a-eca2-4333-b0e4-0b45409c1645-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"6222d35a-eca2-4333-b0e4-0b45409c1645\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 13:58:52 crc kubenswrapper[4840]: I0930 13:58:52.907583 4840 generic.go:334] "Generic (PLEG): container finished" podID="96321113-bb05-4a57-b640-d706993c4520" containerID="90125b546f6eea90269c8f92a5191fca615a408b845afefa0777756279dd6bef" exitCode=0 Sep 30 13:58:52 crc kubenswrapper[4840]: I0930 13:58:52.907653 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mfbwh" event={"ID":"96321113-bb05-4a57-b640-d706993c4520","Type":"ContainerDied","Data":"90125b546f6eea90269c8f92a5191fca615a408b845afefa0777756279dd6bef"} Sep 30 13:58:52 crc kubenswrapper[4840]: I0930 13:58:52.907700 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mfbwh" event={"ID":"96321113-bb05-4a57-b640-d706993c4520","Type":"ContainerStarted","Data":"7bd7c9d17588ff6fb34264a06fa20f8c2d592ce86d4a360ce2ef2c6e173a5d2e"} Sep 30 13:58:52 crc kubenswrapper[4840]: I0930 13:58:52.922940 4840 generic.go:334] "Generic (PLEG): container finished" podID="3c5d8d7f-1aa0-4ae5-98ca-b41d8c9946c1" containerID="2a8f35f539ff7796f447bb7c93da0adac7b72d057e2c2b4773ca1c6785c45a15" exitCode=0 Sep 30 13:58:52 crc kubenswrapper[4840]: I0930 13:58:52.923025 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3c5d8d7f-1aa0-4ae5-98ca-b41d8c9946c1","Type":"ContainerDied","Data":"2a8f35f539ff7796f447bb7c93da0adac7b72d057e2c2b4773ca1c6785c45a15"} Sep 30 13:58:52 crc kubenswrapper[4840]: I0930 13:58:52.936712 4840 generic.go:334] "Generic (PLEG): container finished" podID="3f1c8962-d085-430f-9654-9aa798bc399c" containerID="761f0cb380334134a3c3d0c08581b8420e0880acb55b8526f5a42528399a0f18" exitCode=0 Sep 30 13:58:52 crc kubenswrapper[4840]: I0930 13:58:52.936755 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5vk9z" event={"ID":"3f1c8962-d085-430f-9654-9aa798bc399c","Type":"ContainerDied","Data":"761f0cb380334134a3c3d0c08581b8420e0880acb55b8526f5a42528399a0f18"} Sep 30 13:58:52 crc kubenswrapper[4840]: I0930 13:58:52.993038 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6222d35a-eca2-4333-b0e4-0b45409c1645-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"6222d35a-eca2-4333-b0e4-0b45409c1645\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 13:58:52 crc kubenswrapper[4840]: I0930 13:58:52.993123 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6222d35a-eca2-4333-b0e4-0b45409c1645-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"6222d35a-eca2-4333-b0e4-0b45409c1645\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 13:58:52 crc kubenswrapper[4840]: I0930 13:58:52.993198 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6222d35a-eca2-4333-b0e4-0b45409c1645-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"6222d35a-eca2-4333-b0e4-0b45409c1645\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 13:58:53 crc kubenswrapper[4840]: I0930 13:58:53.019222 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6222d35a-eca2-4333-b0e4-0b45409c1645-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"6222d35a-eca2-4333-b0e4-0b45409c1645\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 13:58:53 crc kubenswrapper[4840]: I0930 13:58:53.139722 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 13:58:53 crc kubenswrapper[4840]: I0930 13:58:53.403950 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Sep 30 13:58:53 crc kubenswrapper[4840]: W0930 13:58:53.430888 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod6222d35a_eca2_4333_b0e4_0b45409c1645.slice/crio-76e62930f7249644471340db61b1077655deee2b15cd4b86719dfa6c39bce903 WatchSource:0}: Error finding container 76e62930f7249644471340db61b1077655deee2b15cd4b86719dfa6c39bce903: Status 404 returned error can't find the container with id 76e62930f7249644471340db61b1077655deee2b15cd4b86719dfa6c39bce903 Sep 30 13:58:53 crc kubenswrapper[4840]: I0930 13:58:53.614300 4840 patch_prober.go:28] interesting pod/router-default-5444994796-s9lx8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 13:58:53 crc kubenswrapper[4840]: [-]has-synced failed: reason withheld Sep 30 13:58:53 crc kubenswrapper[4840]: [+]process-running ok Sep 30 13:58:53 crc kubenswrapper[4840]: healthz check failed Sep 30 13:58:53 crc kubenswrapper[4840]: I0930 13:58:53.614356 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s9lx8" podUID="f0515176-8bf5-46a4-a905-7d308f64449e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 13:58:53 crc kubenswrapper[4840]: I0930 13:58:53.960189 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"6222d35a-eca2-4333-b0e4-0b45409c1645","Type":"ContainerStarted","Data":"76e62930f7249644471340db61b1077655deee2b15cd4b86719dfa6c39bce903"} Sep 30 13:58:54 crc kubenswrapper[4840]: I0930 13:58:54.302034 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 13:58:54 crc kubenswrapper[4840]: I0930 13:58:54.423433 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3c5d8d7f-1aa0-4ae5-98ca-b41d8c9946c1-kube-api-access\") pod \"3c5d8d7f-1aa0-4ae5-98ca-b41d8c9946c1\" (UID: \"3c5d8d7f-1aa0-4ae5-98ca-b41d8c9946c1\") " Sep 30 13:58:54 crc kubenswrapper[4840]: I0930 13:58:54.423538 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3c5d8d7f-1aa0-4ae5-98ca-b41d8c9946c1-kubelet-dir\") pod \"3c5d8d7f-1aa0-4ae5-98ca-b41d8c9946c1\" (UID: \"3c5d8d7f-1aa0-4ae5-98ca-b41d8c9946c1\") " Sep 30 13:58:54 crc kubenswrapper[4840]: I0930 13:58:54.423813 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3c5d8d7f-1aa0-4ae5-98ca-b41d8c9946c1-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "3c5d8d7f-1aa0-4ae5-98ca-b41d8c9946c1" (UID: "3c5d8d7f-1aa0-4ae5-98ca-b41d8c9946c1"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 13:58:54 crc kubenswrapper[4840]: I0930 13:58:54.443925 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c5d8d7f-1aa0-4ae5-98ca-b41d8c9946c1-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "3c5d8d7f-1aa0-4ae5-98ca-b41d8c9946c1" (UID: "3c5d8d7f-1aa0-4ae5-98ca-b41d8c9946c1"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:58:54 crc kubenswrapper[4840]: I0930 13:58:54.524472 4840 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3c5d8d7f-1aa0-4ae5-98ca-b41d8c9946c1-kubelet-dir\") on node \"crc\" DevicePath \"\"" Sep 30 13:58:54 crc kubenswrapper[4840]: I0930 13:58:54.524494 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3c5d8d7f-1aa0-4ae5-98ca-b41d8c9946c1-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 30 13:58:54 crc kubenswrapper[4840]: I0930 13:58:54.615314 4840 patch_prober.go:28] interesting pod/router-default-5444994796-s9lx8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 13:58:54 crc kubenswrapper[4840]: [-]has-synced failed: reason withheld Sep 30 13:58:54 crc kubenswrapper[4840]: [+]process-running ok Sep 30 13:58:54 crc kubenswrapper[4840]: healthz check failed Sep 30 13:58:54 crc kubenswrapper[4840]: I0930 13:58:54.615370 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s9lx8" podUID="f0515176-8bf5-46a4-a905-7d308f64449e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 13:58:54 crc kubenswrapper[4840]: I0930 13:58:54.971094 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"6222d35a-eca2-4333-b0e4-0b45409c1645","Type":"ContainerStarted","Data":"651f89715b154cf2a45069e02372ce84d7eba56b94e5fbb81b3d1c5073501621"} Sep 30 13:58:54 crc kubenswrapper[4840]: I0930 13:58:54.974040 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3c5d8d7f-1aa0-4ae5-98ca-b41d8c9946c1","Type":"ContainerDied","Data":"974c2ce5628eeffaf2764c1ef00ba0f126a42bdadd6130a88116275531de7a2f"} Sep 30 13:58:54 crc kubenswrapper[4840]: I0930 13:58:54.974083 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="974c2ce5628eeffaf2764c1ef00ba0f126a42bdadd6130a88116275531de7a2f" Sep 30 13:58:54 crc kubenswrapper[4840]: I0930 13:58:54.974167 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 13:58:54 crc kubenswrapper[4840]: I0930 13:58:54.984828 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.984811457 podStartE2EDuration="2.984811457s" podCreationTimestamp="2025-09-30 13:58:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:58:54.98415993 +0000 UTC m=+163.613246363" watchObservedRunningTime="2025-09-30 13:58:54.984811457 +0000 UTC m=+163.613897880" Sep 30 13:58:55 crc kubenswrapper[4840]: I0930 13:58:55.576325 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" Sep 30 13:58:55 crc kubenswrapper[4840]: I0930 13:58:55.581497 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-lk4jg" Sep 30 13:58:55 crc kubenswrapper[4840]: I0930 13:58:55.616022 4840 patch_prober.go:28] interesting pod/router-default-5444994796-s9lx8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 13:58:55 crc kubenswrapper[4840]: [-]has-synced failed: reason withheld Sep 30 13:58:55 crc kubenswrapper[4840]: [+]process-running ok Sep 30 13:58:55 crc kubenswrapper[4840]: healthz check failed Sep 30 13:58:55 crc kubenswrapper[4840]: I0930 13:58:55.616080 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s9lx8" podUID="f0515176-8bf5-46a4-a905-7d308f64449e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 13:58:55 crc kubenswrapper[4840]: I0930 13:58:55.858137 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-68gcw" Sep 30 13:58:55 crc kubenswrapper[4840]: I0930 13:58:55.982843 4840 generic.go:334] "Generic (PLEG): container finished" podID="6222d35a-eca2-4333-b0e4-0b45409c1645" containerID="651f89715b154cf2a45069e02372ce84d7eba56b94e5fbb81b3d1c5073501621" exitCode=0 Sep 30 13:58:55 crc kubenswrapper[4840]: I0930 13:58:55.982908 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"6222d35a-eca2-4333-b0e4-0b45409c1645","Type":"ContainerDied","Data":"651f89715b154cf2a45069e02372ce84d7eba56b94e5fbb81b3d1c5073501621"} Sep 30 13:58:56 crc kubenswrapper[4840]: I0930 13:58:56.615136 4840 patch_prober.go:28] interesting pod/router-default-5444994796-s9lx8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 13:58:56 crc kubenswrapper[4840]: [-]has-synced failed: reason withheld Sep 30 13:58:56 crc kubenswrapper[4840]: [+]process-running ok Sep 30 13:58:56 crc kubenswrapper[4840]: healthz check failed Sep 30 13:58:56 crc kubenswrapper[4840]: I0930 13:58:56.615253 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s9lx8" podUID="f0515176-8bf5-46a4-a905-7d308f64449e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 13:58:57 crc kubenswrapper[4840]: I0930 13:58:57.615129 4840 patch_prober.go:28] interesting pod/router-default-5444994796-s9lx8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 13:58:57 crc kubenswrapper[4840]: [-]has-synced failed: reason withheld Sep 30 13:58:57 crc kubenswrapper[4840]: [+]process-running ok Sep 30 13:58:57 crc kubenswrapper[4840]: healthz check failed Sep 30 13:58:57 crc kubenswrapper[4840]: I0930 13:58:57.615206 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s9lx8" podUID="f0515176-8bf5-46a4-a905-7d308f64449e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 13:58:57 crc kubenswrapper[4840]: I0930 13:58:57.894000 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1491f559-bc12-4afd-a40c-4eaa40d920a8-metrics-certs\") pod \"network-metrics-daemon-4gp5x\" (UID: \"1491f559-bc12-4afd-a40c-4eaa40d920a8\") " pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:58:57 crc kubenswrapper[4840]: I0930 13:58:57.899539 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1491f559-bc12-4afd-a40c-4eaa40d920a8-metrics-certs\") pod \"network-metrics-daemon-4gp5x\" (UID: \"1491f559-bc12-4afd-a40c-4eaa40d920a8\") " pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:58:58 crc kubenswrapper[4840]: I0930 13:58:58.059290 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4gp5x" Sep 30 13:58:58 crc kubenswrapper[4840]: I0930 13:58:58.614269 4840 patch_prober.go:28] interesting pod/router-default-5444994796-s9lx8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 13:58:58 crc kubenswrapper[4840]: [-]has-synced failed: reason withheld Sep 30 13:58:58 crc kubenswrapper[4840]: [+]process-running ok Sep 30 13:58:58 crc kubenswrapper[4840]: healthz check failed Sep 30 13:58:58 crc kubenswrapper[4840]: I0930 13:58:58.614324 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s9lx8" podUID="f0515176-8bf5-46a4-a905-7d308f64449e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 13:58:59 crc kubenswrapper[4840]: I0930 13:58:59.102121 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 13:58:59 crc kubenswrapper[4840]: I0930 13:58:59.209201 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6222d35a-eca2-4333-b0e4-0b45409c1645-kube-api-access\") pod \"6222d35a-eca2-4333-b0e4-0b45409c1645\" (UID: \"6222d35a-eca2-4333-b0e4-0b45409c1645\") " Sep 30 13:58:59 crc kubenswrapper[4840]: I0930 13:58:59.209321 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6222d35a-eca2-4333-b0e4-0b45409c1645-kubelet-dir\") pod \"6222d35a-eca2-4333-b0e4-0b45409c1645\" (UID: \"6222d35a-eca2-4333-b0e4-0b45409c1645\") " Sep 30 13:58:59 crc kubenswrapper[4840]: I0930 13:58:59.209793 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6222d35a-eca2-4333-b0e4-0b45409c1645-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "6222d35a-eca2-4333-b0e4-0b45409c1645" (UID: "6222d35a-eca2-4333-b0e4-0b45409c1645"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 13:58:59 crc kubenswrapper[4840]: I0930 13:58:59.212725 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6222d35a-eca2-4333-b0e4-0b45409c1645-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "6222d35a-eca2-4333-b0e4-0b45409c1645" (UID: "6222d35a-eca2-4333-b0e4-0b45409c1645"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:58:59 crc kubenswrapper[4840]: I0930 13:58:59.310371 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6222d35a-eca2-4333-b0e4-0b45409c1645-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 30 13:58:59 crc kubenswrapper[4840]: I0930 13:58:59.310407 4840 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6222d35a-eca2-4333-b0e4-0b45409c1645-kubelet-dir\") on node \"crc\" DevicePath \"\"" Sep 30 13:58:59 crc kubenswrapper[4840]: I0930 13:58:59.613944 4840 patch_prober.go:28] interesting pod/router-default-5444994796-s9lx8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 13:58:59 crc kubenswrapper[4840]: [-]has-synced failed: reason withheld Sep 30 13:58:59 crc kubenswrapper[4840]: [+]process-running ok Sep 30 13:58:59 crc kubenswrapper[4840]: healthz check failed Sep 30 13:58:59 crc kubenswrapper[4840]: I0930 13:58:59.614008 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s9lx8" podUID="f0515176-8bf5-46a4-a905-7d308f64449e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 13:59:00 crc kubenswrapper[4840]: I0930 13:59:00.007380 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"6222d35a-eca2-4333-b0e4-0b45409c1645","Type":"ContainerDied","Data":"76e62930f7249644471340db61b1077655deee2b15cd4b86719dfa6c39bce903"} Sep 30 13:59:00 crc kubenswrapper[4840]: I0930 13:59:00.007425 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76e62930f7249644471340db61b1077655deee2b15cd4b86719dfa6c39bce903" Sep 30 13:59:00 crc kubenswrapper[4840]: I0930 13:59:00.007517 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 13:59:00 crc kubenswrapper[4840]: I0930 13:59:00.134634 4840 patch_prober.go:28] interesting pod/downloads-7954f5f757-kdwqh container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Sep 30 13:59:00 crc kubenswrapper[4840]: I0930 13:59:00.134700 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-kdwqh" podUID="dde0d17e-6bd0-4acb-b8b8-a8b0bc988a6e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Sep 30 13:59:00 crc kubenswrapper[4840]: I0930 13:59:00.135068 4840 patch_prober.go:28] interesting pod/downloads-7954f5f757-kdwqh container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Sep 30 13:59:00 crc kubenswrapper[4840]: I0930 13:59:00.135099 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-kdwqh" podUID="dde0d17e-6bd0-4acb-b8b8-a8b0bc988a6e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Sep 30 13:59:00 crc kubenswrapper[4840]: I0930 13:59:00.346013 4840 patch_prober.go:28] interesting pod/console-f9d7485db-qjxnf container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.6:8443/health\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Sep 30 13:59:00 crc kubenswrapper[4840]: I0930 13:59:00.346410 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-qjxnf" podUID="9dd57b46-77fc-4b8a-9fc1-a50084165914" containerName="console" probeResult="failure" output="Get \"https://10.217.0.6:8443/health\": dial tcp 10.217.0.6:8443: connect: connection refused" Sep 30 13:59:00 crc kubenswrapper[4840]: I0930 13:59:00.613536 4840 patch_prober.go:28] interesting pod/router-default-5444994796-s9lx8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 13:59:00 crc kubenswrapper[4840]: [-]has-synced failed: reason withheld Sep 30 13:59:00 crc kubenswrapper[4840]: [+]process-running ok Sep 30 13:59:00 crc kubenswrapper[4840]: healthz check failed Sep 30 13:59:00 crc kubenswrapper[4840]: I0930 13:59:00.613621 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s9lx8" podUID="f0515176-8bf5-46a4-a905-7d308f64449e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 13:59:01 crc kubenswrapper[4840]: I0930 13:59:01.617488 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-s9lx8" Sep 30 13:59:01 crc kubenswrapper[4840]: I0930 13:59:01.621865 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-s9lx8" Sep 30 13:59:09 crc kubenswrapper[4840]: I0930 13:59:09.697783 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 13:59:10 crc kubenswrapper[4840]: E0930 13:59:10.034679 4840 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Sep 30 13:59:10 crc kubenswrapper[4840]: E0930 13:59:10.034876 4840 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6t49m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-z895r_openshift-marketplace(5d587bdc-9a81-4e7c-9df2-d638d43f5e0d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 30 13:59:10 crc kubenswrapper[4840]: E0930 13:59:10.036301 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-z895r" podUID="5d587bdc-9a81-4e7c-9df2-d638d43f5e0d" Sep 30 13:59:10 crc kubenswrapper[4840]: I0930 13:59:10.138208 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-kdwqh" Sep 30 13:59:10 crc kubenswrapper[4840]: I0930 13:59:10.350199 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-qjxnf" Sep 30 13:59:10 crc kubenswrapper[4840]: I0930 13:59:10.355837 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-qjxnf" Sep 30 13:59:11 crc kubenswrapper[4840]: E0930 13:59:11.056766 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-z895r" podUID="5d587bdc-9a81-4e7c-9df2-d638d43f5e0d" Sep 30 13:59:12 crc kubenswrapper[4840]: E0930 13:59:12.969654 4840 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Sep 30 13:59:12 crc kubenswrapper[4840]: E0930 13:59:12.969914 4840 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bt794,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-qxvn2_openshift-marketplace(3c2b4d42-9935-4e56-972f-cdd95d5a0e86): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 30 13:59:12 crc kubenswrapper[4840]: E0930 13:59:12.971203 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-qxvn2" podUID="3c2b4d42-9935-4e56-972f-cdd95d5a0e86" Sep 30 13:59:19 crc kubenswrapper[4840]: E0930 13:59:19.923726 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-qxvn2" podUID="3c2b4d42-9935-4e56-972f-cdd95d5a0e86" Sep 30 13:59:20 crc kubenswrapper[4840]: E0930 13:59:20.032596 4840 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Sep 30 13:59:20 crc kubenswrapper[4840]: E0930 13:59:20.032949 4840 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lzr4v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-dd72s_openshift-marketplace(fcd2c899-747f-42ed-b1c4-cf6b70b5b230): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 30 13:59:20 crc kubenswrapper[4840]: E0930 13:59:20.034099 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-dd72s" podUID="fcd2c899-747f-42ed-b1c4-cf6b70b5b230" Sep 30 13:59:20 crc kubenswrapper[4840]: E0930 13:59:20.035352 4840 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Sep 30 13:59:20 crc kubenswrapper[4840]: E0930 13:59:20.035483 4840 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p9ljk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-xn8ts_openshift-marketplace(796c97de-5fd5-4ffa-886b-3ffa2ead1da3): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 30 13:59:20 crc kubenswrapper[4840]: E0930 13:59:20.036649 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-xn8ts" podUID="796c97de-5fd5-4ffa-886b-3ffa2ead1da3" Sep 30 13:59:20 crc kubenswrapper[4840]: E0930 13:59:20.061145 4840 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Sep 30 13:59:20 crc kubenswrapper[4840]: E0930 13:59:20.061294 4840 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fddm9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-rdflm_openshift-marketplace(a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 30 13:59:20 crc kubenswrapper[4840]: E0930 13:59:20.062444 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-rdflm" podUID="a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4" Sep 30 13:59:20 crc kubenswrapper[4840]: I0930 13:59:20.306161 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-4gp5x"] Sep 30 13:59:21 crc kubenswrapper[4840]: I0930 13:59:21.099964 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9sxdn" Sep 30 13:59:21 crc kubenswrapper[4840]: I0930 13:59:21.872267 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 13:59:21 crc kubenswrapper[4840]: I0930 13:59:21.872330 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 13:59:22 crc kubenswrapper[4840]: E0930 13:59:22.693895 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-rdflm" podUID="a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4" Sep 30 13:59:22 crc kubenswrapper[4840]: E0930 13:59:22.693949 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-xn8ts" podUID="796c97de-5fd5-4ffa-886b-3ffa2ead1da3" Sep 30 13:59:22 crc kubenswrapper[4840]: E0930 13:59:22.694401 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-dd72s" podUID="fcd2c899-747f-42ed-b1c4-cf6b70b5b230" Sep 30 13:59:22 crc kubenswrapper[4840]: E0930 13:59:22.716444 4840 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Sep 30 13:59:22 crc kubenswrapper[4840]: E0930 13:59:22.716703 4840 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x64f5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-5vk9z_openshift-marketplace(3f1c8962-d085-430f-9654-9aa798bc399c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 30 13:59:22 crc kubenswrapper[4840]: E0930 13:59:22.717209 4840 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Sep 30 13:59:22 crc kubenswrapper[4840]: E0930 13:59:22.717290 4840 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vrsbc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-mfbwh_openshift-marketplace(96321113-bb05-4a57-b640-d706993c4520): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 30 13:59:22 crc kubenswrapper[4840]: E0930 13:59:22.718345 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-5vk9z" podUID="3f1c8962-d085-430f-9654-9aa798bc399c" Sep 30 13:59:22 crc kubenswrapper[4840]: E0930 13:59:22.718355 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-mfbwh" podUID="96321113-bb05-4a57-b640-d706993c4520" Sep 30 13:59:23 crc kubenswrapper[4840]: I0930 13:59:23.150840 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-4gp5x" event={"ID":"1491f559-bc12-4afd-a40c-4eaa40d920a8","Type":"ContainerStarted","Data":"35fa21ac15ec7de42f6b389ce40c5e360d080d52a353cc4475cbf92977b1f38a"} Sep 30 13:59:23 crc kubenswrapper[4840]: I0930 13:59:23.150893 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-4gp5x" event={"ID":"1491f559-bc12-4afd-a40c-4eaa40d920a8","Type":"ContainerStarted","Data":"ce3b7b019dae192b66c5a3c6409ddbd9774b109929d0c0ff392ae2411507a5cb"} Sep 30 13:59:23 crc kubenswrapper[4840]: I0930 13:59:23.150903 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-4gp5x" event={"ID":"1491f559-bc12-4afd-a40c-4eaa40d920a8","Type":"ContainerStarted","Data":"dd95293636ab5a4710c8e5de2a16f3d3ac6ec5635d734d34d5addbf25a5c243b"} Sep 30 13:59:23 crc kubenswrapper[4840]: I0930 13:59:23.152899 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tc2fx" event={"ID":"c97b5b98-9a09-4c3b-9bd0-5a5a8671c479","Type":"ContainerStarted","Data":"8d3a35c625e0dee7fbe9ee695b32c4d5c35612188e2e65fd036fcb781d3abdaf"} Sep 30 13:59:23 crc kubenswrapper[4840]: E0930 13:59:23.154889 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-mfbwh" podUID="96321113-bb05-4a57-b640-d706993c4520" Sep 30 13:59:23 crc kubenswrapper[4840]: E0930 13:59:23.155249 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-5vk9z" podUID="3f1c8962-d085-430f-9654-9aa798bc399c" Sep 30 13:59:23 crc kubenswrapper[4840]: I0930 13:59:23.185250 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-4gp5x" podStartSLOduration=169.185225314 podStartE2EDuration="2m49.185225314s" podCreationTimestamp="2025-09-30 13:56:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 13:59:23.166827288 +0000 UTC m=+191.795913751" watchObservedRunningTime="2025-09-30 13:59:23.185225314 +0000 UTC m=+191.814311757" Sep 30 13:59:24 crc kubenswrapper[4840]: I0930 13:59:24.159438 4840 generic.go:334] "Generic (PLEG): container finished" podID="c97b5b98-9a09-4c3b-9bd0-5a5a8671c479" containerID="8d3a35c625e0dee7fbe9ee695b32c4d5c35612188e2e65fd036fcb781d3abdaf" exitCode=0 Sep 30 13:59:24 crc kubenswrapper[4840]: I0930 13:59:24.159581 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tc2fx" event={"ID":"c97b5b98-9a09-4c3b-9bd0-5a5a8671c479","Type":"ContainerDied","Data":"8d3a35c625e0dee7fbe9ee695b32c4d5c35612188e2e65fd036fcb781d3abdaf"} Sep 30 13:59:24 crc kubenswrapper[4840]: I0930 13:59:24.163252 4840 generic.go:334] "Generic (PLEG): container finished" podID="5d587bdc-9a81-4e7c-9df2-d638d43f5e0d" containerID="11b5a464fb1cdc01fdad7819b80af6a830c396a49fe2a970dd8011998936cb20" exitCode=0 Sep 30 13:59:24 crc kubenswrapper[4840]: I0930 13:59:24.163351 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z895r" event={"ID":"5d587bdc-9a81-4e7c-9df2-d638d43f5e0d","Type":"ContainerDied","Data":"11b5a464fb1cdc01fdad7819b80af6a830c396a49fe2a970dd8011998936cb20"} Sep 30 13:59:25 crc kubenswrapper[4840]: I0930 13:59:25.173490 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tc2fx" event={"ID":"c97b5b98-9a09-4c3b-9bd0-5a5a8671c479","Type":"ContainerStarted","Data":"13fae0bb486eb5bb58d70a75d2ed58249ead39d9548a8d6c09fe2c422ddd815c"} Sep 30 13:59:25 crc kubenswrapper[4840]: I0930 13:59:25.177335 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z895r" event={"ID":"5d587bdc-9a81-4e7c-9df2-d638d43f5e0d","Type":"ContainerStarted","Data":"f18006d0f42332874ec1886fc12d62142fbc2cc6ec63aa08d8041df79e5f8f5c"} Sep 30 13:59:25 crc kubenswrapper[4840]: I0930 13:59:25.214772 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tc2fx" podStartSLOduration=3.404257274 podStartE2EDuration="38.21475004s" podCreationTimestamp="2025-09-30 13:58:47 +0000 UTC" firstStartedPulling="2025-09-30 13:58:49.822276464 +0000 UTC m=+158.451362887" lastFinishedPulling="2025-09-30 13:59:24.63276923 +0000 UTC m=+193.261855653" observedRunningTime="2025-09-30 13:59:25.197905945 +0000 UTC m=+193.826992358" watchObservedRunningTime="2025-09-30 13:59:25.21475004 +0000 UTC m=+193.843836473" Sep 30 13:59:25 crc kubenswrapper[4840]: I0930 13:59:25.217646 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-z895r" podStartSLOduration=2.51958515 podStartE2EDuration="36.217633487s" podCreationTimestamp="2025-09-30 13:58:49 +0000 UTC" firstStartedPulling="2025-09-30 13:58:50.869879629 +0000 UTC m=+159.498966052" lastFinishedPulling="2025-09-30 13:59:24.567927966 +0000 UTC m=+193.197014389" observedRunningTime="2025-09-30 13:59:25.213538688 +0000 UTC m=+193.842625111" watchObservedRunningTime="2025-09-30 13:59:25.217633487 +0000 UTC m=+193.846719920" Sep 30 13:59:27 crc kubenswrapper[4840]: I0930 13:59:27.135287 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 13:59:27 crc kubenswrapper[4840]: I0930 13:59:27.983288 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tc2fx" Sep 30 13:59:27 crc kubenswrapper[4840]: I0930 13:59:27.983613 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tc2fx" Sep 30 13:59:28 crc kubenswrapper[4840]: I0930 13:59:28.124019 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tc2fx" Sep 30 13:59:29 crc kubenswrapper[4840]: I0930 13:59:29.271782 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tc2fx" Sep 30 13:59:29 crc kubenswrapper[4840]: I0930 13:59:29.775365 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-z895r" Sep 30 13:59:29 crc kubenswrapper[4840]: I0930 13:59:29.775758 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-z895r" Sep 30 13:59:29 crc kubenswrapper[4840]: I0930 13:59:29.812499 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-z895r" Sep 30 13:59:30 crc kubenswrapper[4840]: I0930 13:59:30.244848 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-z895r" Sep 30 13:59:34 crc kubenswrapper[4840]: I0930 13:59:34.237504 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qxvn2" event={"ID":"3c2b4d42-9935-4e56-972f-cdd95d5a0e86","Type":"ContainerStarted","Data":"8a4822dab2e81685d61813fc0e1f6505f9349dee991207ba566c3443e8ea414f"} Sep 30 13:59:35 crc kubenswrapper[4840]: I0930 13:59:35.245464 4840 generic.go:334] "Generic (PLEG): container finished" podID="796c97de-5fd5-4ffa-886b-3ffa2ead1da3" containerID="1d81f6d3c3ea1c9ba9b761132d16bb87cd1034ffc9144d32d49d3c69df42ede3" exitCode=0 Sep 30 13:59:35 crc kubenswrapper[4840]: I0930 13:59:35.245542 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xn8ts" event={"ID":"796c97de-5fd5-4ffa-886b-3ffa2ead1da3","Type":"ContainerDied","Data":"1d81f6d3c3ea1c9ba9b761132d16bb87cd1034ffc9144d32d49d3c69df42ede3"} Sep 30 13:59:35 crc kubenswrapper[4840]: I0930 13:59:35.249207 4840 generic.go:334] "Generic (PLEG): container finished" podID="3c2b4d42-9935-4e56-972f-cdd95d5a0e86" containerID="8a4822dab2e81685d61813fc0e1f6505f9349dee991207ba566c3443e8ea414f" exitCode=0 Sep 30 13:59:35 crc kubenswrapper[4840]: I0930 13:59:35.249295 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qxvn2" event={"ID":"3c2b4d42-9935-4e56-972f-cdd95d5a0e86","Type":"ContainerDied","Data":"8a4822dab2e81685d61813fc0e1f6505f9349dee991207ba566c3443e8ea414f"} Sep 30 13:59:35 crc kubenswrapper[4840]: I0930 13:59:35.251371 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rdflm" event={"ID":"a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4","Type":"ContainerStarted","Data":"826a50371992cd7995ff94ca3c17153c28b19fa3f8d13fd89df91bf1aeb60fa5"} Sep 30 13:59:36 crc kubenswrapper[4840]: I0930 13:59:36.258700 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xn8ts" event={"ID":"796c97de-5fd5-4ffa-886b-3ffa2ead1da3","Type":"ContainerStarted","Data":"768d643bc9f5c989ea2359f4a4955b6cfb32da6484c3a1c20b9eec57ca51371d"} Sep 30 13:59:36 crc kubenswrapper[4840]: I0930 13:59:36.261321 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qxvn2" event={"ID":"3c2b4d42-9935-4e56-972f-cdd95d5a0e86","Type":"ContainerStarted","Data":"cac37e0372fb79b04eda0d36f6395e9124b5455e61b468361d9bad4e1ea4ba4e"} Sep 30 13:59:36 crc kubenswrapper[4840]: I0930 13:59:36.262837 4840 generic.go:334] "Generic (PLEG): container finished" podID="a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4" containerID="826a50371992cd7995ff94ca3c17153c28b19fa3f8d13fd89df91bf1aeb60fa5" exitCode=0 Sep 30 13:59:36 crc kubenswrapper[4840]: I0930 13:59:36.262899 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rdflm" event={"ID":"a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4","Type":"ContainerDied","Data":"826a50371992cd7995ff94ca3c17153c28b19fa3f8d13fd89df91bf1aeb60fa5"} Sep 30 13:59:36 crc kubenswrapper[4840]: I0930 13:59:36.264527 4840 generic.go:334] "Generic (PLEG): container finished" podID="fcd2c899-747f-42ed-b1c4-cf6b70b5b230" containerID="489466a2a8c9709e8510f7b9e6aa06c2f72b8c452532f83b1b9a35b62c56d3b4" exitCode=0 Sep 30 13:59:36 crc kubenswrapper[4840]: I0930 13:59:36.264594 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dd72s" event={"ID":"fcd2c899-747f-42ed-b1c4-cf6b70b5b230","Type":"ContainerDied","Data":"489466a2a8c9709e8510f7b9e6aa06c2f72b8c452532f83b1b9a35b62c56d3b4"} Sep 30 13:59:36 crc kubenswrapper[4840]: I0930 13:59:36.278490 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xn8ts" podStartSLOduration=2.27873892 podStartE2EDuration="47.278474198s" podCreationTimestamp="2025-09-30 13:58:49 +0000 UTC" firstStartedPulling="2025-09-30 13:58:50.866994903 +0000 UTC m=+159.496081326" lastFinishedPulling="2025-09-30 13:59:35.866730181 +0000 UTC m=+204.495816604" observedRunningTime="2025-09-30 13:59:36.277710148 +0000 UTC m=+204.906796571" watchObservedRunningTime="2025-09-30 13:59:36.278474198 +0000 UTC m=+204.907560621" Sep 30 13:59:36 crc kubenswrapper[4840]: I0930 13:59:36.295832 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qxvn2" podStartSLOduration=3.385732417 podStartE2EDuration="49.295809565s" podCreationTimestamp="2025-09-30 13:58:47 +0000 UTC" firstStartedPulling="2025-09-30 13:58:49.813660306 +0000 UTC m=+158.442746729" lastFinishedPulling="2025-09-30 13:59:35.723737454 +0000 UTC m=+204.352823877" observedRunningTime="2025-09-30 13:59:36.294913251 +0000 UTC m=+204.923999694" watchObservedRunningTime="2025-09-30 13:59:36.295809565 +0000 UTC m=+204.924895988" Sep 30 13:59:37 crc kubenswrapper[4840]: I0930 13:59:37.271592 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rdflm" event={"ID":"a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4","Type":"ContainerStarted","Data":"48c5fdee1cf43506529a5c18e0a7fadd452d817c35d157057f87effb14ad0b18"} Sep 30 13:59:37 crc kubenswrapper[4840]: I0930 13:59:37.273634 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dd72s" event={"ID":"fcd2c899-747f-42ed-b1c4-cf6b70b5b230","Type":"ContainerStarted","Data":"ddc1cc6562ef08391a109ad89b4f2fdbb37b8a5aec3c689ece7d31ea609380cb"} Sep 30 13:59:37 crc kubenswrapper[4840]: I0930 13:59:37.300467 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rdflm" podStartSLOduration=3.334322764 podStartE2EDuration="50.300452479s" podCreationTimestamp="2025-09-30 13:58:47 +0000 UTC" firstStartedPulling="2025-09-30 13:58:49.826531207 +0000 UTC m=+158.455617630" lastFinishedPulling="2025-09-30 13:59:36.792660922 +0000 UTC m=+205.421747345" observedRunningTime="2025-09-30 13:59:37.298374175 +0000 UTC m=+205.927460608" watchObservedRunningTime="2025-09-30 13:59:37.300452479 +0000 UTC m=+205.929538902" Sep 30 13:59:37 crc kubenswrapper[4840]: I0930 13:59:37.321486 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dd72s" podStartSLOduration=2.426720618 podStartE2EDuration="49.321467383s" podCreationTimestamp="2025-09-30 13:58:48 +0000 UTC" firstStartedPulling="2025-09-30 13:58:49.848490897 +0000 UTC m=+158.477577320" lastFinishedPulling="2025-09-30 13:59:36.743237662 +0000 UTC m=+205.372324085" observedRunningTime="2025-09-30 13:59:37.319684986 +0000 UTC m=+205.948771409" watchObservedRunningTime="2025-09-30 13:59:37.321467383 +0000 UTC m=+205.950553806" Sep 30 13:59:37 crc kubenswrapper[4840]: I0930 13:59:37.872843 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qxvn2" Sep 30 13:59:37 crc kubenswrapper[4840]: I0930 13:59:37.873980 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qxvn2" Sep 30 13:59:37 crc kubenswrapper[4840]: I0930 13:59:37.921000 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qxvn2" Sep 30 13:59:38 crc kubenswrapper[4840]: I0930 13:59:38.215961 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rdflm" Sep 30 13:59:38 crc kubenswrapper[4840]: I0930 13:59:38.216273 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rdflm" Sep 30 13:59:38 crc kubenswrapper[4840]: I0930 13:59:38.456309 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dd72s" Sep 30 13:59:38 crc kubenswrapper[4840]: I0930 13:59:38.456371 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dd72s" Sep 30 13:59:39 crc kubenswrapper[4840]: I0930 13:59:39.259495 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-rdflm" podUID="a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4" containerName="registry-server" probeResult="failure" output=< Sep 30 13:59:39 crc kubenswrapper[4840]: timeout: failed to connect service ":50051" within 1s Sep 30 13:59:39 crc kubenswrapper[4840]: > Sep 30 13:59:39 crc kubenswrapper[4840]: I0930 13:59:39.284422 4840 generic.go:334] "Generic (PLEG): container finished" podID="3f1c8962-d085-430f-9654-9aa798bc399c" containerID="645ed392836c63be996642ef20a1a081037ffb00d13e2eae9c24b844c0c9f701" exitCode=0 Sep 30 13:59:39 crc kubenswrapper[4840]: I0930 13:59:39.284504 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5vk9z" event={"ID":"3f1c8962-d085-430f-9654-9aa798bc399c","Type":"ContainerDied","Data":"645ed392836c63be996642ef20a1a081037ffb00d13e2eae9c24b844c0c9f701"} Sep 30 13:59:39 crc kubenswrapper[4840]: I0930 13:59:39.492354 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-dd72s" podUID="fcd2c899-747f-42ed-b1c4-cf6b70b5b230" containerName="registry-server" probeResult="failure" output=< Sep 30 13:59:39 crc kubenswrapper[4840]: timeout: failed to connect service ":50051" within 1s Sep 30 13:59:39 crc kubenswrapper[4840]: > Sep 30 13:59:40 crc kubenswrapper[4840]: I0930 13:59:40.173153 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xn8ts" Sep 30 13:59:40 crc kubenswrapper[4840]: I0930 13:59:40.173533 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xn8ts" Sep 30 13:59:40 crc kubenswrapper[4840]: I0930 13:59:40.213411 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xn8ts" Sep 30 13:59:40 crc kubenswrapper[4840]: I0930 13:59:40.290655 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mfbwh" event={"ID":"96321113-bb05-4a57-b640-d706993c4520","Type":"ContainerStarted","Data":"e86bcc387126a26b548e9228fd034c0ee53cc3260f9475ab30387d4ce35b5c16"} Sep 30 13:59:41 crc kubenswrapper[4840]: I0930 13:59:41.298210 4840 generic.go:334] "Generic (PLEG): container finished" podID="96321113-bb05-4a57-b640-d706993c4520" containerID="e86bcc387126a26b548e9228fd034c0ee53cc3260f9475ab30387d4ce35b5c16" exitCode=0 Sep 30 13:59:41 crc kubenswrapper[4840]: I0930 13:59:41.298260 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mfbwh" event={"ID":"96321113-bb05-4a57-b640-d706993c4520","Type":"ContainerDied","Data":"e86bcc387126a26b548e9228fd034c0ee53cc3260f9475ab30387d4ce35b5c16"} Sep 30 13:59:47 crc kubenswrapper[4840]: I0930 13:59:47.916791 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qxvn2" Sep 30 13:59:48 crc kubenswrapper[4840]: I0930 13:59:48.254909 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rdflm" Sep 30 13:59:48 crc kubenswrapper[4840]: I0930 13:59:48.290682 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rdflm" Sep 30 13:59:48 crc kubenswrapper[4840]: I0930 13:59:48.496750 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dd72s" Sep 30 13:59:48 crc kubenswrapper[4840]: I0930 13:59:48.543643 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dd72s" Sep 30 13:59:50 crc kubenswrapper[4840]: I0930 13:59:50.100896 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rdflm"] Sep 30 13:59:50 crc kubenswrapper[4840]: I0930 13:59:50.101903 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rdflm" podUID="a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4" containerName="registry-server" containerID="cri-o://48c5fdee1cf43506529a5c18e0a7fadd452d817c35d157057f87effb14ad0b18" gracePeriod=2 Sep 30 13:59:50 crc kubenswrapper[4840]: I0930 13:59:50.211081 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xn8ts" Sep 30 13:59:50 crc kubenswrapper[4840]: I0930 13:59:50.302665 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dd72s"] Sep 30 13:59:50 crc kubenswrapper[4840]: I0930 13:59:50.348704 4840 generic.go:334] "Generic (PLEG): container finished" podID="a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4" containerID="48c5fdee1cf43506529a5c18e0a7fadd452d817c35d157057f87effb14ad0b18" exitCode=0 Sep 30 13:59:50 crc kubenswrapper[4840]: I0930 13:59:50.348765 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rdflm" event={"ID":"a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4","Type":"ContainerDied","Data":"48c5fdee1cf43506529a5c18e0a7fadd452d817c35d157057f87effb14ad0b18"} Sep 30 13:59:50 crc kubenswrapper[4840]: I0930 13:59:50.349049 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dd72s" podUID="fcd2c899-747f-42ed-b1c4-cf6b70b5b230" containerName="registry-server" containerID="cri-o://ddc1cc6562ef08391a109ad89b4f2fdbb37b8a5aec3c689ece7d31ea609380cb" gracePeriod=2 Sep 30 13:59:50 crc kubenswrapper[4840]: I0930 13:59:50.844488 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rdflm" Sep 30 13:59:50 crc kubenswrapper[4840]: I0930 13:59:50.887161 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4-utilities\") pod \"a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4\" (UID: \"a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4\") " Sep 30 13:59:50 crc kubenswrapper[4840]: I0930 13:59:50.887232 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4-catalog-content\") pod \"a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4\" (UID: \"a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4\") " Sep 30 13:59:50 crc kubenswrapper[4840]: I0930 13:59:50.887260 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fddm9\" (UniqueName: \"kubernetes.io/projected/a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4-kube-api-access-fddm9\") pod \"a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4\" (UID: \"a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4\") " Sep 30 13:59:50 crc kubenswrapper[4840]: I0930 13:59:50.888089 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4-utilities" (OuterVolumeSpecName: "utilities") pod "a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4" (UID: "a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 13:59:50 crc kubenswrapper[4840]: I0930 13:59:50.894138 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4-kube-api-access-fddm9" (OuterVolumeSpecName: "kube-api-access-fddm9") pod "a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4" (UID: "a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4"). InnerVolumeSpecName "kube-api-access-fddm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:59:50 crc kubenswrapper[4840]: I0930 13:59:50.937489 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4" (UID: "a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 13:59:50 crc kubenswrapper[4840]: I0930 13:59:50.989492 4840 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 13:59:50 crc kubenswrapper[4840]: I0930 13:59:50.989536 4840 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 13:59:50 crc kubenswrapper[4840]: I0930 13:59:50.989550 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fddm9\" (UniqueName: \"kubernetes.io/projected/a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4-kube-api-access-fddm9\") on node \"crc\" DevicePath \"\"" Sep 30 13:59:51 crc kubenswrapper[4840]: I0930 13:59:51.366817 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rdflm" Sep 30 13:59:51 crc kubenswrapper[4840]: I0930 13:59:51.366843 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rdflm" event={"ID":"a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4","Type":"ContainerDied","Data":"e8c21dcafd69f1fd2c21f53ee7d5d0f95b6ff7fff97a4739f6154b7747e2bea7"} Sep 30 13:59:51 crc kubenswrapper[4840]: I0930 13:59:51.366937 4840 scope.go:117] "RemoveContainer" containerID="48c5fdee1cf43506529a5c18e0a7fadd452d817c35d157057f87effb14ad0b18" Sep 30 13:59:51 crc kubenswrapper[4840]: I0930 13:59:51.374615 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dd72s" event={"ID":"fcd2c899-747f-42ed-b1c4-cf6b70b5b230","Type":"ContainerDied","Data":"ddc1cc6562ef08391a109ad89b4f2fdbb37b8a5aec3c689ece7d31ea609380cb"} Sep 30 13:59:51 crc kubenswrapper[4840]: I0930 13:59:51.374620 4840 generic.go:334] "Generic (PLEG): container finished" podID="fcd2c899-747f-42ed-b1c4-cf6b70b5b230" containerID="ddc1cc6562ef08391a109ad89b4f2fdbb37b8a5aec3c689ece7d31ea609380cb" exitCode=0 Sep 30 13:59:51 crc kubenswrapper[4840]: I0930 13:59:51.392084 4840 scope.go:117] "RemoveContainer" containerID="826a50371992cd7995ff94ca3c17153c28b19fa3f8d13fd89df91bf1aeb60fa5" Sep 30 13:59:51 crc kubenswrapper[4840]: I0930 13:59:51.400556 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rdflm"] Sep 30 13:59:51 crc kubenswrapper[4840]: I0930 13:59:51.403993 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rdflm"] Sep 30 13:59:51 crc kubenswrapper[4840]: I0930 13:59:51.418335 4840 scope.go:117] "RemoveContainer" containerID="1edd4f8f22d64d076189d4949a8d935a45780a20d367f23ac4e9f35ef6d668c2" Sep 30 13:59:51 crc kubenswrapper[4840]: I0930 13:59:51.587745 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dd72s" Sep 30 13:59:51 crc kubenswrapper[4840]: I0930 13:59:51.699045 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzr4v\" (UniqueName: \"kubernetes.io/projected/fcd2c899-747f-42ed-b1c4-cf6b70b5b230-kube-api-access-lzr4v\") pod \"fcd2c899-747f-42ed-b1c4-cf6b70b5b230\" (UID: \"fcd2c899-747f-42ed-b1c4-cf6b70b5b230\") " Sep 30 13:59:51 crc kubenswrapper[4840]: I0930 13:59:51.699185 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcd2c899-747f-42ed-b1c4-cf6b70b5b230-catalog-content\") pod \"fcd2c899-747f-42ed-b1c4-cf6b70b5b230\" (UID: \"fcd2c899-747f-42ed-b1c4-cf6b70b5b230\") " Sep 30 13:59:51 crc kubenswrapper[4840]: I0930 13:59:51.699305 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcd2c899-747f-42ed-b1c4-cf6b70b5b230-utilities\") pod \"fcd2c899-747f-42ed-b1c4-cf6b70b5b230\" (UID: \"fcd2c899-747f-42ed-b1c4-cf6b70b5b230\") " Sep 30 13:59:51 crc kubenswrapper[4840]: I0930 13:59:51.700111 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fcd2c899-747f-42ed-b1c4-cf6b70b5b230-utilities" (OuterVolumeSpecName: "utilities") pod "fcd2c899-747f-42ed-b1c4-cf6b70b5b230" (UID: "fcd2c899-747f-42ed-b1c4-cf6b70b5b230"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 13:59:51 crc kubenswrapper[4840]: I0930 13:59:51.704357 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcd2c899-747f-42ed-b1c4-cf6b70b5b230-kube-api-access-lzr4v" (OuterVolumeSpecName: "kube-api-access-lzr4v") pod "fcd2c899-747f-42ed-b1c4-cf6b70b5b230" (UID: "fcd2c899-747f-42ed-b1c4-cf6b70b5b230"). InnerVolumeSpecName "kube-api-access-lzr4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:59:51 crc kubenswrapper[4840]: I0930 13:59:51.753238 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fcd2c899-747f-42ed-b1c4-cf6b70b5b230-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fcd2c899-747f-42ed-b1c4-cf6b70b5b230" (UID: "fcd2c899-747f-42ed-b1c4-cf6b70b5b230"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 13:59:51 crc kubenswrapper[4840]: I0930 13:59:51.800742 4840 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcd2c899-747f-42ed-b1c4-cf6b70b5b230-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 13:59:51 crc kubenswrapper[4840]: I0930 13:59:51.800783 4840 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcd2c899-747f-42ed-b1c4-cf6b70b5b230-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 13:59:51 crc kubenswrapper[4840]: I0930 13:59:51.800797 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzr4v\" (UniqueName: \"kubernetes.io/projected/fcd2c899-747f-42ed-b1c4-cf6b70b5b230-kube-api-access-lzr4v\") on node \"crc\" DevicePath \"\"" Sep 30 13:59:51 crc kubenswrapper[4840]: I0930 13:59:51.872232 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 13:59:51 crc kubenswrapper[4840]: I0930 13:59:51.872294 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 13:59:51 crc kubenswrapper[4840]: I0930 13:59:51.872337 4840 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-747gk" Sep 30 13:59:51 crc kubenswrapper[4840]: I0930 13:59:51.872797 4840 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd"} pod="openshift-machine-config-operator/machine-config-daemon-747gk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 13:59:51 crc kubenswrapper[4840]: I0930 13:59:51.872898 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" containerID="cri-o://ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd" gracePeriod=600 Sep 30 13:59:52 crc kubenswrapper[4840]: I0930 13:59:52.134765 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4" path="/var/lib/kubelet/pods/a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4/volumes" Sep 30 13:59:52 crc kubenswrapper[4840]: I0930 13:59:52.388410 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5vk9z" event={"ID":"3f1c8962-d085-430f-9654-9aa798bc399c","Type":"ContainerStarted","Data":"4b0cf2afc6e7946683643afb3dcf6ed4670240ed4be90e19061174d54b826154"} Sep 30 13:59:52 crc kubenswrapper[4840]: I0930 13:59:52.391133 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mfbwh" event={"ID":"96321113-bb05-4a57-b640-d706993c4520","Type":"ContainerStarted","Data":"8e7434869cb556379db813a2fb78144274b96ac0f9f9e5f6928826250efee536"} Sep 30 13:59:52 crc kubenswrapper[4840]: I0930 13:59:52.393377 4840 generic.go:334] "Generic (PLEG): container finished" podID="10e8b890-7f20-4a36-8e03-898620cf599a" containerID="ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd" exitCode=0 Sep 30 13:59:52 crc kubenswrapper[4840]: I0930 13:59:52.393442 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" event={"ID":"10e8b890-7f20-4a36-8e03-898620cf599a","Type":"ContainerDied","Data":"ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd"} Sep 30 13:59:52 crc kubenswrapper[4840]: I0930 13:59:52.396502 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dd72s" event={"ID":"fcd2c899-747f-42ed-b1c4-cf6b70b5b230","Type":"ContainerDied","Data":"035c917febad21cbcf2d1a650107529376c542b8f40b7dc2e5c6b10b1c62d8c2"} Sep 30 13:59:52 crc kubenswrapper[4840]: I0930 13:59:52.396535 4840 scope.go:117] "RemoveContainer" containerID="ddc1cc6562ef08391a109ad89b4f2fdbb37b8a5aec3c689ece7d31ea609380cb" Sep 30 13:59:52 crc kubenswrapper[4840]: I0930 13:59:52.396598 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dd72s" Sep 30 13:59:52 crc kubenswrapper[4840]: I0930 13:59:52.410732 4840 scope.go:117] "RemoveContainer" containerID="489466a2a8c9709e8510f7b9e6aa06c2f72b8c452532f83b1b9a35b62c56d3b4" Sep 30 13:59:52 crc kubenswrapper[4840]: I0930 13:59:52.414971 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5vk9z" podStartSLOduration=4.78347713 podStartE2EDuration="1m2.414952611s" podCreationTimestamp="2025-09-30 13:58:50 +0000 UTC" firstStartedPulling="2025-09-30 13:58:52.939504795 +0000 UTC m=+161.568591218" lastFinishedPulling="2025-09-30 13:59:50.570980266 +0000 UTC m=+219.200066699" observedRunningTime="2025-09-30 13:59:52.412515197 +0000 UTC m=+221.041601650" watchObservedRunningTime="2025-09-30 13:59:52.414952611 +0000 UTC m=+221.044039034" Sep 30 13:59:52 crc kubenswrapper[4840]: I0930 13:59:52.427138 4840 scope.go:117] "RemoveContainer" containerID="7ae770494d9a9fac05d23e9accc3db4865620fe0c704e646230a1e69b61473d8" Sep 30 13:59:52 crc kubenswrapper[4840]: I0930 13:59:52.427842 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mfbwh" podStartSLOduration=3.768221386 podStartE2EDuration="1m1.42782254s" podCreationTimestamp="2025-09-30 13:58:51 +0000 UTC" firstStartedPulling="2025-09-30 13:58:52.910354335 +0000 UTC m=+161.539440758" lastFinishedPulling="2025-09-30 13:59:50.569955489 +0000 UTC m=+219.199041912" observedRunningTime="2025-09-30 13:59:52.42668103 +0000 UTC m=+221.055767463" watchObservedRunningTime="2025-09-30 13:59:52.42782254 +0000 UTC m=+221.056908973" Sep 30 13:59:52 crc kubenswrapper[4840]: I0930 13:59:52.440997 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dd72s"] Sep 30 13:59:52 crc kubenswrapper[4840]: I0930 13:59:52.444438 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dd72s"] Sep 30 13:59:52 crc kubenswrapper[4840]: I0930 13:59:52.501100 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xn8ts"] Sep 30 13:59:52 crc kubenswrapper[4840]: I0930 13:59:52.501417 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xn8ts" podUID="796c97de-5fd5-4ffa-886b-3ffa2ead1da3" containerName="registry-server" containerID="cri-o://768d643bc9f5c989ea2359f4a4955b6cfb32da6484c3a1c20b9eec57ca51371d" gracePeriod=2 Sep 30 13:59:53 crc kubenswrapper[4840]: I0930 13:59:53.405288 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" event={"ID":"10e8b890-7f20-4a36-8e03-898620cf599a","Type":"ContainerStarted","Data":"e98140cf872382a187fd61e08bfb9d9a34afdc7cbc64b4d05d64d64c86bc9322"} Sep 30 13:59:54 crc kubenswrapper[4840]: I0930 13:59:54.142178 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fcd2c899-747f-42ed-b1c4-cf6b70b5b230" path="/var/lib/kubelet/pods/fcd2c899-747f-42ed-b1c4-cf6b70b5b230/volumes" Sep 30 13:59:55 crc kubenswrapper[4840]: I0930 13:59:55.422148 4840 generic.go:334] "Generic (PLEG): container finished" podID="796c97de-5fd5-4ffa-886b-3ffa2ead1da3" containerID="768d643bc9f5c989ea2359f4a4955b6cfb32da6484c3a1c20b9eec57ca51371d" exitCode=0 Sep 30 13:59:55 crc kubenswrapper[4840]: I0930 13:59:55.422205 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xn8ts" event={"ID":"796c97de-5fd5-4ffa-886b-3ffa2ead1da3","Type":"ContainerDied","Data":"768d643bc9f5c989ea2359f4a4955b6cfb32da6484c3a1c20b9eec57ca51371d"} Sep 30 13:59:55 crc kubenswrapper[4840]: I0930 13:59:55.747574 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xn8ts" Sep 30 13:59:55 crc kubenswrapper[4840]: I0930 13:59:55.850487 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/796c97de-5fd5-4ffa-886b-3ffa2ead1da3-catalog-content\") pod \"796c97de-5fd5-4ffa-886b-3ffa2ead1da3\" (UID: \"796c97de-5fd5-4ffa-886b-3ffa2ead1da3\") " Sep 30 13:59:55 crc kubenswrapper[4840]: I0930 13:59:55.850896 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/796c97de-5fd5-4ffa-886b-3ffa2ead1da3-utilities\") pod \"796c97de-5fd5-4ffa-886b-3ffa2ead1da3\" (UID: \"796c97de-5fd5-4ffa-886b-3ffa2ead1da3\") " Sep 30 13:59:55 crc kubenswrapper[4840]: I0930 13:59:55.850964 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p9ljk\" (UniqueName: \"kubernetes.io/projected/796c97de-5fd5-4ffa-886b-3ffa2ead1da3-kube-api-access-p9ljk\") pod \"796c97de-5fd5-4ffa-886b-3ffa2ead1da3\" (UID: \"796c97de-5fd5-4ffa-886b-3ffa2ead1da3\") " Sep 30 13:59:55 crc kubenswrapper[4840]: I0930 13:59:55.851868 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/796c97de-5fd5-4ffa-886b-3ffa2ead1da3-utilities" (OuterVolumeSpecName: "utilities") pod "796c97de-5fd5-4ffa-886b-3ffa2ead1da3" (UID: "796c97de-5fd5-4ffa-886b-3ffa2ead1da3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 13:59:55 crc kubenswrapper[4840]: I0930 13:59:55.857445 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/796c97de-5fd5-4ffa-886b-3ffa2ead1da3-kube-api-access-p9ljk" (OuterVolumeSpecName: "kube-api-access-p9ljk") pod "796c97de-5fd5-4ffa-886b-3ffa2ead1da3" (UID: "796c97de-5fd5-4ffa-886b-3ffa2ead1da3"). InnerVolumeSpecName "kube-api-access-p9ljk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 13:59:55 crc kubenswrapper[4840]: I0930 13:59:55.952558 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p9ljk\" (UniqueName: \"kubernetes.io/projected/796c97de-5fd5-4ffa-886b-3ffa2ead1da3-kube-api-access-p9ljk\") on node \"crc\" DevicePath \"\"" Sep 30 13:59:55 crc kubenswrapper[4840]: I0930 13:59:55.952605 4840 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/796c97de-5fd5-4ffa-886b-3ffa2ead1da3-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 13:59:55 crc kubenswrapper[4840]: I0930 13:59:55.957964 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/796c97de-5fd5-4ffa-886b-3ffa2ead1da3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "796c97de-5fd5-4ffa-886b-3ffa2ead1da3" (UID: "796c97de-5fd5-4ffa-886b-3ffa2ead1da3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 13:59:56 crc kubenswrapper[4840]: I0930 13:59:56.054175 4840 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/796c97de-5fd5-4ffa-886b-3ffa2ead1da3-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 13:59:56 crc kubenswrapper[4840]: I0930 13:59:56.429998 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xn8ts" event={"ID":"796c97de-5fd5-4ffa-886b-3ffa2ead1da3","Type":"ContainerDied","Data":"2ae45e4e3399b314482af5674bcf5bf6fe6499c8c6499db52fad547803d6bbca"} Sep 30 13:59:56 crc kubenswrapper[4840]: I0930 13:59:56.430045 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xn8ts" Sep 30 13:59:56 crc kubenswrapper[4840]: I0930 13:59:56.430049 4840 scope.go:117] "RemoveContainer" containerID="768d643bc9f5c989ea2359f4a4955b6cfb32da6484c3a1c20b9eec57ca51371d" Sep 30 13:59:56 crc kubenswrapper[4840]: I0930 13:59:56.448498 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xn8ts"] Sep 30 13:59:56 crc kubenswrapper[4840]: I0930 13:59:56.451914 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xn8ts"] Sep 30 13:59:56 crc kubenswrapper[4840]: I0930 13:59:56.458924 4840 scope.go:117] "RemoveContainer" containerID="1d81f6d3c3ea1c9ba9b761132d16bb87cd1034ffc9144d32d49d3c69df42ede3" Sep 30 13:59:56 crc kubenswrapper[4840]: I0930 13:59:56.474432 4840 scope.go:117] "RemoveContainer" containerID="f4c3cb6ee1ced67ef507f0dd198b81d6dee92f4c8c43786c75d9cd43e7723236" Sep 30 13:59:58 crc kubenswrapper[4840]: I0930 13:59:58.124218 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="796c97de-5fd5-4ffa-886b-3ffa2ead1da3" path="/var/lib/kubelet/pods/796c97de-5fd5-4ffa-886b-3ffa2ead1da3/volumes" Sep 30 14:00:00 crc kubenswrapper[4840]: I0930 14:00:00.134779 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320680-vtwpz"] Sep 30 14:00:00 crc kubenswrapper[4840]: E0930 14:00:00.135023 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="796c97de-5fd5-4ffa-886b-3ffa2ead1da3" containerName="extract-content" Sep 30 14:00:00 crc kubenswrapper[4840]: I0930 14:00:00.135039 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="796c97de-5fd5-4ffa-886b-3ffa2ead1da3" containerName="extract-content" Sep 30 14:00:00 crc kubenswrapper[4840]: E0930 14:00:00.135052 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6222d35a-eca2-4333-b0e4-0b45409c1645" containerName="pruner" Sep 30 14:00:00 crc kubenswrapper[4840]: I0930 14:00:00.135060 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="6222d35a-eca2-4333-b0e4-0b45409c1645" containerName="pruner" Sep 30 14:00:00 crc kubenswrapper[4840]: E0930 14:00:00.135072 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c5d8d7f-1aa0-4ae5-98ca-b41d8c9946c1" containerName="pruner" Sep 30 14:00:00 crc kubenswrapper[4840]: I0930 14:00:00.135080 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c5d8d7f-1aa0-4ae5-98ca-b41d8c9946c1" containerName="pruner" Sep 30 14:00:00 crc kubenswrapper[4840]: E0930 14:00:00.135092 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="796c97de-5fd5-4ffa-886b-3ffa2ead1da3" containerName="registry-server" Sep 30 14:00:00 crc kubenswrapper[4840]: I0930 14:00:00.135099 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="796c97de-5fd5-4ffa-886b-3ffa2ead1da3" containerName="registry-server" Sep 30 14:00:00 crc kubenswrapper[4840]: E0930 14:00:00.135111 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4" containerName="registry-server" Sep 30 14:00:00 crc kubenswrapper[4840]: I0930 14:00:00.135118 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4" containerName="registry-server" Sep 30 14:00:00 crc kubenswrapper[4840]: E0930 14:00:00.135133 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4" containerName="extract-content" Sep 30 14:00:00 crc kubenswrapper[4840]: I0930 14:00:00.135140 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4" containerName="extract-content" Sep 30 14:00:00 crc kubenswrapper[4840]: E0930 14:00:00.135150 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcd2c899-747f-42ed-b1c4-cf6b70b5b230" containerName="extract-content" Sep 30 14:00:00 crc kubenswrapper[4840]: I0930 14:00:00.135157 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcd2c899-747f-42ed-b1c4-cf6b70b5b230" containerName="extract-content" Sep 30 14:00:00 crc kubenswrapper[4840]: E0930 14:00:00.135167 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4" containerName="extract-utilities" Sep 30 14:00:00 crc kubenswrapper[4840]: I0930 14:00:00.135175 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4" containerName="extract-utilities" Sep 30 14:00:00 crc kubenswrapper[4840]: E0930 14:00:00.135183 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcd2c899-747f-42ed-b1c4-cf6b70b5b230" containerName="extract-utilities" Sep 30 14:00:00 crc kubenswrapper[4840]: I0930 14:00:00.135189 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcd2c899-747f-42ed-b1c4-cf6b70b5b230" containerName="extract-utilities" Sep 30 14:00:00 crc kubenswrapper[4840]: E0930 14:00:00.135199 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcd2c899-747f-42ed-b1c4-cf6b70b5b230" containerName="registry-server" Sep 30 14:00:00 crc kubenswrapper[4840]: I0930 14:00:00.135206 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcd2c899-747f-42ed-b1c4-cf6b70b5b230" containerName="registry-server" Sep 30 14:00:00 crc kubenswrapper[4840]: E0930 14:00:00.135219 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="796c97de-5fd5-4ffa-886b-3ffa2ead1da3" containerName="extract-utilities" Sep 30 14:00:00 crc kubenswrapper[4840]: I0930 14:00:00.135226 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="796c97de-5fd5-4ffa-886b-3ffa2ead1da3" containerName="extract-utilities" Sep 30 14:00:00 crc kubenswrapper[4840]: I0930 14:00:00.135333 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c5d8d7f-1aa0-4ae5-98ca-b41d8c9946c1" containerName="pruner" Sep 30 14:00:00 crc kubenswrapper[4840]: I0930 14:00:00.135347 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="6222d35a-eca2-4333-b0e4-0b45409c1645" containerName="pruner" Sep 30 14:00:00 crc kubenswrapper[4840]: I0930 14:00:00.135357 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="796c97de-5fd5-4ffa-886b-3ffa2ead1da3" containerName="registry-server" Sep 30 14:00:00 crc kubenswrapper[4840]: I0930 14:00:00.135366 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcd2c899-747f-42ed-b1c4-cf6b70b5b230" containerName="registry-server" Sep 30 14:00:00 crc kubenswrapper[4840]: I0930 14:00:00.135375 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2ec4007-7bc3-43dd-bbe5-b172bb7b30e4" containerName="registry-server" Sep 30 14:00:00 crc kubenswrapper[4840]: I0930 14:00:00.135883 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320680-vtwpz" Sep 30 14:00:00 crc kubenswrapper[4840]: I0930 14:00:00.137994 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 30 14:00:00 crc kubenswrapper[4840]: I0930 14:00:00.137994 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 30 14:00:00 crc kubenswrapper[4840]: I0930 14:00:00.144056 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320680-vtwpz"] Sep 30 14:00:00 crc kubenswrapper[4840]: I0930 14:00:00.212760 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kx9wq\" (UniqueName: \"kubernetes.io/projected/556ac8c1-39c4-4a47-a542-6881f953ef8f-kube-api-access-kx9wq\") pod \"collect-profiles-29320680-vtwpz\" (UID: \"556ac8c1-39c4-4a47-a542-6881f953ef8f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320680-vtwpz" Sep 30 14:00:00 crc kubenswrapper[4840]: I0930 14:00:00.212863 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/556ac8c1-39c4-4a47-a542-6881f953ef8f-config-volume\") pod \"collect-profiles-29320680-vtwpz\" (UID: \"556ac8c1-39c4-4a47-a542-6881f953ef8f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320680-vtwpz" Sep 30 14:00:00 crc kubenswrapper[4840]: I0930 14:00:00.212917 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/556ac8c1-39c4-4a47-a542-6881f953ef8f-secret-volume\") pod \"collect-profiles-29320680-vtwpz\" (UID: \"556ac8c1-39c4-4a47-a542-6881f953ef8f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320680-vtwpz" Sep 30 14:00:00 crc kubenswrapper[4840]: I0930 14:00:00.265033 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ncblm"] Sep 30 14:00:00 crc kubenswrapper[4840]: I0930 14:00:00.313907 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/556ac8c1-39c4-4a47-a542-6881f953ef8f-secret-volume\") pod \"collect-profiles-29320680-vtwpz\" (UID: \"556ac8c1-39c4-4a47-a542-6881f953ef8f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320680-vtwpz" Sep 30 14:00:00 crc kubenswrapper[4840]: I0930 14:00:00.314325 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kx9wq\" (UniqueName: \"kubernetes.io/projected/556ac8c1-39c4-4a47-a542-6881f953ef8f-kube-api-access-kx9wq\") pod \"collect-profiles-29320680-vtwpz\" (UID: \"556ac8c1-39c4-4a47-a542-6881f953ef8f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320680-vtwpz" Sep 30 14:00:00 crc kubenswrapper[4840]: I0930 14:00:00.314395 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/556ac8c1-39c4-4a47-a542-6881f953ef8f-config-volume\") pod \"collect-profiles-29320680-vtwpz\" (UID: \"556ac8c1-39c4-4a47-a542-6881f953ef8f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320680-vtwpz" Sep 30 14:00:00 crc kubenswrapper[4840]: I0930 14:00:00.315629 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/556ac8c1-39c4-4a47-a542-6881f953ef8f-config-volume\") pod \"collect-profiles-29320680-vtwpz\" (UID: \"556ac8c1-39c4-4a47-a542-6881f953ef8f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320680-vtwpz" Sep 30 14:00:00 crc kubenswrapper[4840]: I0930 14:00:00.320189 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/556ac8c1-39c4-4a47-a542-6881f953ef8f-secret-volume\") pod \"collect-profiles-29320680-vtwpz\" (UID: \"556ac8c1-39c4-4a47-a542-6881f953ef8f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320680-vtwpz" Sep 30 14:00:00 crc kubenswrapper[4840]: I0930 14:00:00.333972 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kx9wq\" (UniqueName: \"kubernetes.io/projected/556ac8c1-39c4-4a47-a542-6881f953ef8f-kube-api-access-kx9wq\") pod \"collect-profiles-29320680-vtwpz\" (UID: \"556ac8c1-39c4-4a47-a542-6881f953ef8f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320680-vtwpz" Sep 30 14:00:00 crc kubenswrapper[4840]: I0930 14:00:00.450363 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320680-vtwpz" Sep 30 14:00:00 crc kubenswrapper[4840]: I0930 14:00:00.912049 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320680-vtwpz"] Sep 30 14:00:01 crc kubenswrapper[4840]: I0930 14:00:01.167344 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5vk9z" Sep 30 14:00:01 crc kubenswrapper[4840]: I0930 14:00:01.167702 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5vk9z" Sep 30 14:00:01 crc kubenswrapper[4840]: I0930 14:00:01.209398 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5vk9z" Sep 30 14:00:01 crc kubenswrapper[4840]: I0930 14:00:01.457467 4840 generic.go:334] "Generic (PLEG): container finished" podID="556ac8c1-39c4-4a47-a542-6881f953ef8f" containerID="87134cf6d6900bd5d20a83e4fe6e39464405a640d3334de5c538f352c5bd5dce" exitCode=0 Sep 30 14:00:01 crc kubenswrapper[4840]: I0930 14:00:01.457592 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320680-vtwpz" event={"ID":"556ac8c1-39c4-4a47-a542-6881f953ef8f","Type":"ContainerDied","Data":"87134cf6d6900bd5d20a83e4fe6e39464405a640d3334de5c538f352c5bd5dce"} Sep 30 14:00:01 crc kubenswrapper[4840]: I0930 14:00:01.457677 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320680-vtwpz" event={"ID":"556ac8c1-39c4-4a47-a542-6881f953ef8f","Type":"ContainerStarted","Data":"4a52f81fdd010a7a431d617059a7d69cac6bb0bb73c9241dbb3fd76aae48f3d7"} Sep 30 14:00:01 crc kubenswrapper[4840]: I0930 14:00:01.501066 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5vk9z" Sep 30 14:00:01 crc kubenswrapper[4840]: I0930 14:00:01.560979 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mfbwh" Sep 30 14:00:01 crc kubenswrapper[4840]: I0930 14:00:01.561473 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mfbwh" Sep 30 14:00:01 crc kubenswrapper[4840]: I0930 14:00:01.595277 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mfbwh" Sep 30 14:00:02 crc kubenswrapper[4840]: I0930 14:00:02.519117 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mfbwh" Sep 30 14:00:02 crc kubenswrapper[4840]: I0930 14:00:02.759411 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320680-vtwpz" Sep 30 14:00:02 crc kubenswrapper[4840]: I0930 14:00:02.853643 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/556ac8c1-39c4-4a47-a542-6881f953ef8f-config-volume\") pod \"556ac8c1-39c4-4a47-a542-6881f953ef8f\" (UID: \"556ac8c1-39c4-4a47-a542-6881f953ef8f\") " Sep 30 14:00:02 crc kubenswrapper[4840]: I0930 14:00:02.853716 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kx9wq\" (UniqueName: \"kubernetes.io/projected/556ac8c1-39c4-4a47-a542-6881f953ef8f-kube-api-access-kx9wq\") pod \"556ac8c1-39c4-4a47-a542-6881f953ef8f\" (UID: \"556ac8c1-39c4-4a47-a542-6881f953ef8f\") " Sep 30 14:00:02 crc kubenswrapper[4840]: I0930 14:00:02.853833 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/556ac8c1-39c4-4a47-a542-6881f953ef8f-secret-volume\") pod \"556ac8c1-39c4-4a47-a542-6881f953ef8f\" (UID: \"556ac8c1-39c4-4a47-a542-6881f953ef8f\") " Sep 30 14:00:02 crc kubenswrapper[4840]: I0930 14:00:02.854605 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/556ac8c1-39c4-4a47-a542-6881f953ef8f-config-volume" (OuterVolumeSpecName: "config-volume") pod "556ac8c1-39c4-4a47-a542-6881f953ef8f" (UID: "556ac8c1-39c4-4a47-a542-6881f953ef8f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:00:02 crc kubenswrapper[4840]: I0930 14:00:02.866748 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/556ac8c1-39c4-4a47-a542-6881f953ef8f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "556ac8c1-39c4-4a47-a542-6881f953ef8f" (UID: "556ac8c1-39c4-4a47-a542-6881f953ef8f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:00:02 crc kubenswrapper[4840]: I0930 14:00:02.866788 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/556ac8c1-39c4-4a47-a542-6881f953ef8f-kube-api-access-kx9wq" (OuterVolumeSpecName: "kube-api-access-kx9wq") pod "556ac8c1-39c4-4a47-a542-6881f953ef8f" (UID: "556ac8c1-39c4-4a47-a542-6881f953ef8f"). InnerVolumeSpecName "kube-api-access-kx9wq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:00:02 crc kubenswrapper[4840]: I0930 14:00:02.955289 4840 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/556ac8c1-39c4-4a47-a542-6881f953ef8f-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 30 14:00:02 crc kubenswrapper[4840]: I0930 14:00:02.955330 4840 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/556ac8c1-39c4-4a47-a542-6881f953ef8f-config-volume\") on node \"crc\" DevicePath \"\"" Sep 30 14:00:02 crc kubenswrapper[4840]: I0930 14:00:02.955344 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kx9wq\" (UniqueName: \"kubernetes.io/projected/556ac8c1-39c4-4a47-a542-6881f953ef8f-kube-api-access-kx9wq\") on node \"crc\" DevicePath \"\"" Sep 30 14:00:03 crc kubenswrapper[4840]: I0930 14:00:03.477269 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320680-vtwpz" Sep 30 14:00:03 crc kubenswrapper[4840]: I0930 14:00:03.477325 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320680-vtwpz" event={"ID":"556ac8c1-39c4-4a47-a542-6881f953ef8f","Type":"ContainerDied","Data":"4a52f81fdd010a7a431d617059a7d69cac6bb0bb73c9241dbb3fd76aae48f3d7"} Sep 30 14:00:03 crc kubenswrapper[4840]: I0930 14:00:03.477350 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a52f81fdd010a7a431d617059a7d69cac6bb0bb73c9241dbb3fd76aae48f3d7" Sep 30 14:00:03 crc kubenswrapper[4840]: I0930 14:00:03.698523 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mfbwh"] Sep 30 14:00:05 crc kubenswrapper[4840]: I0930 14:00:05.485625 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mfbwh" podUID="96321113-bb05-4a57-b640-d706993c4520" containerName="registry-server" containerID="cri-o://8e7434869cb556379db813a2fb78144274b96ac0f9f9e5f6928826250efee536" gracePeriod=2 Sep 30 14:00:06 crc kubenswrapper[4840]: I0930 14:00:06.495133 4840 generic.go:334] "Generic (PLEG): container finished" podID="96321113-bb05-4a57-b640-d706993c4520" containerID="8e7434869cb556379db813a2fb78144274b96ac0f9f9e5f6928826250efee536" exitCode=0 Sep 30 14:00:06 crc kubenswrapper[4840]: I0930 14:00:06.495405 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mfbwh" event={"ID":"96321113-bb05-4a57-b640-d706993c4520","Type":"ContainerDied","Data":"8e7434869cb556379db813a2fb78144274b96ac0f9f9e5f6928826250efee536"} Sep 30 14:00:06 crc kubenswrapper[4840]: I0930 14:00:06.671472 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mfbwh" Sep 30 14:00:06 crc kubenswrapper[4840]: I0930 14:00:06.801086 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrsbc\" (UniqueName: \"kubernetes.io/projected/96321113-bb05-4a57-b640-d706993c4520-kube-api-access-vrsbc\") pod \"96321113-bb05-4a57-b640-d706993c4520\" (UID: \"96321113-bb05-4a57-b640-d706993c4520\") " Sep 30 14:00:06 crc kubenswrapper[4840]: I0930 14:00:06.801169 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96321113-bb05-4a57-b640-d706993c4520-utilities\") pod \"96321113-bb05-4a57-b640-d706993c4520\" (UID: \"96321113-bb05-4a57-b640-d706993c4520\") " Sep 30 14:00:06 crc kubenswrapper[4840]: I0930 14:00:06.801248 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96321113-bb05-4a57-b640-d706993c4520-catalog-content\") pod \"96321113-bb05-4a57-b640-d706993c4520\" (UID: \"96321113-bb05-4a57-b640-d706993c4520\") " Sep 30 14:00:06 crc kubenswrapper[4840]: I0930 14:00:06.802199 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96321113-bb05-4a57-b640-d706993c4520-utilities" (OuterVolumeSpecName: "utilities") pod "96321113-bb05-4a57-b640-d706993c4520" (UID: "96321113-bb05-4a57-b640-d706993c4520"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:00:06 crc kubenswrapper[4840]: I0930 14:00:06.806480 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96321113-bb05-4a57-b640-d706993c4520-kube-api-access-vrsbc" (OuterVolumeSpecName: "kube-api-access-vrsbc") pod "96321113-bb05-4a57-b640-d706993c4520" (UID: "96321113-bb05-4a57-b640-d706993c4520"). InnerVolumeSpecName "kube-api-access-vrsbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:00:06 crc kubenswrapper[4840]: I0930 14:00:06.902477 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrsbc\" (UniqueName: \"kubernetes.io/projected/96321113-bb05-4a57-b640-d706993c4520-kube-api-access-vrsbc\") on node \"crc\" DevicePath \"\"" Sep 30 14:00:06 crc kubenswrapper[4840]: I0930 14:00:06.902516 4840 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96321113-bb05-4a57-b640-d706993c4520-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:00:07 crc kubenswrapper[4840]: I0930 14:00:07.086017 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96321113-bb05-4a57-b640-d706993c4520-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "96321113-bb05-4a57-b640-d706993c4520" (UID: "96321113-bb05-4a57-b640-d706993c4520"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:00:07 crc kubenswrapper[4840]: I0930 14:00:07.105003 4840 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96321113-bb05-4a57-b640-d706993c4520-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:00:07 crc kubenswrapper[4840]: I0930 14:00:07.504000 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mfbwh" event={"ID":"96321113-bb05-4a57-b640-d706993c4520","Type":"ContainerDied","Data":"7bd7c9d17588ff6fb34264a06fa20f8c2d592ce86d4a360ce2ef2c6e173a5d2e"} Sep 30 14:00:07 crc kubenswrapper[4840]: I0930 14:00:07.504065 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mfbwh" Sep 30 14:00:07 crc kubenswrapper[4840]: I0930 14:00:07.504334 4840 scope.go:117] "RemoveContainer" containerID="8e7434869cb556379db813a2fb78144274b96ac0f9f9e5f6928826250efee536" Sep 30 14:00:07 crc kubenswrapper[4840]: I0930 14:00:07.521420 4840 scope.go:117] "RemoveContainer" containerID="e86bcc387126a26b548e9228fd034c0ee53cc3260f9475ab30387d4ce35b5c16" Sep 30 14:00:07 crc kubenswrapper[4840]: I0930 14:00:07.530930 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mfbwh"] Sep 30 14:00:07 crc kubenswrapper[4840]: I0930 14:00:07.533370 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mfbwh"] Sep 30 14:00:07 crc kubenswrapper[4840]: I0930 14:00:07.542591 4840 scope.go:117] "RemoveContainer" containerID="90125b546f6eea90269c8f92a5191fca615a408b845afefa0777756279dd6bef" Sep 30 14:00:08 crc kubenswrapper[4840]: I0930 14:00:08.125004 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96321113-bb05-4a57-b640-d706993c4520" path="/var/lib/kubelet/pods/96321113-bb05-4a57-b640-d706993c4520/volumes" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.288733 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" podUID="c1b9a905-662c-4e2a-8066-ee8d75e5d45b" containerName="oauth-openshift" containerID="cri-o://a8aca6442d2827c689d94cb212fdbd982789b44b0113a53ff16a3cb3e0364162" gracePeriod=15 Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.600683 4840 generic.go:334] "Generic (PLEG): container finished" podID="c1b9a905-662c-4e2a-8066-ee8d75e5d45b" containerID="a8aca6442d2827c689d94cb212fdbd982789b44b0113a53ff16a3cb3e0364162" exitCode=0 Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.600787 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" event={"ID":"c1b9a905-662c-4e2a-8066-ee8d75e5d45b","Type":"ContainerDied","Data":"a8aca6442d2827c689d94cb212fdbd982789b44b0113a53ff16a3cb3e0364162"} Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.680090 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.708045 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh"] Sep 30 14:00:25 crc kubenswrapper[4840]: E0930 14:00:25.708254 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96321113-bb05-4a57-b640-d706993c4520" containerName="registry-server" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.708269 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="96321113-bb05-4a57-b640-d706993c4520" containerName="registry-server" Sep 30 14:00:25 crc kubenswrapper[4840]: E0930 14:00:25.708281 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96321113-bb05-4a57-b640-d706993c4520" containerName="extract-content" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.708288 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="96321113-bb05-4a57-b640-d706993c4520" containerName="extract-content" Sep 30 14:00:25 crc kubenswrapper[4840]: E0930 14:00:25.708301 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96321113-bb05-4a57-b640-d706993c4520" containerName="extract-utilities" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.708308 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="96321113-bb05-4a57-b640-d706993c4520" containerName="extract-utilities" Sep 30 14:00:25 crc kubenswrapper[4840]: E0930 14:00:25.708326 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1b9a905-662c-4e2a-8066-ee8d75e5d45b" containerName="oauth-openshift" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.708334 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1b9a905-662c-4e2a-8066-ee8d75e5d45b" containerName="oauth-openshift" Sep 30 14:00:25 crc kubenswrapper[4840]: E0930 14:00:25.708343 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="556ac8c1-39c4-4a47-a542-6881f953ef8f" containerName="collect-profiles" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.708352 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="556ac8c1-39c4-4a47-a542-6881f953ef8f" containerName="collect-profiles" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.708455 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="556ac8c1-39c4-4a47-a542-6881f953ef8f" containerName="collect-profiles" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.708471 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1b9a905-662c-4e2a-8066-ee8d75e5d45b" containerName="oauth-openshift" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.708482 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="96321113-bb05-4a57-b640-d706993c4520" containerName="registry-server" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.708922 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.725419 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh"] Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.740916 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-audit-policies\") pod \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.740978 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-ocp-branding-template\") pod \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.741003 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-router-certs\") pod \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.741026 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-cliconfig\") pod \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.741043 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-trusted-ca-bundle\") pod \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.741064 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-user-template-provider-selection\") pod \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.741107 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-serving-cert\") pod \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.741123 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-user-idp-0-file-data\") pod \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.741143 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-session\") pod \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.741163 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-audit-dir\") pod \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.741217 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-user-template-error\") pod \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.741243 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-service-ca\") pod \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.741257 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-user-template-login\") pod \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.741277 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmrwg\" (UniqueName: \"kubernetes.io/projected/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-kube-api-access-qmrwg\") pod \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\" (UID: \"c1b9a905-662c-4e2a-8066-ee8d75e5d45b\") " Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.741684 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "c1b9a905-662c-4e2a-8066-ee8d75e5d45b" (UID: "c1b9a905-662c-4e2a-8066-ee8d75e5d45b"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.742233 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "c1b9a905-662c-4e2a-8066-ee8d75e5d45b" (UID: "c1b9a905-662c-4e2a-8066-ee8d75e5d45b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.742941 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "c1b9a905-662c-4e2a-8066-ee8d75e5d45b" (UID: "c1b9a905-662c-4e2a-8066-ee8d75e5d45b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.743082 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "c1b9a905-662c-4e2a-8066-ee8d75e5d45b" (UID: "c1b9a905-662c-4e2a-8066-ee8d75e5d45b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.743441 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "c1b9a905-662c-4e2a-8066-ee8d75e5d45b" (UID: "c1b9a905-662c-4e2a-8066-ee8d75e5d45b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.747226 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "c1b9a905-662c-4e2a-8066-ee8d75e5d45b" (UID: "c1b9a905-662c-4e2a-8066-ee8d75e5d45b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.747625 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-kube-api-access-qmrwg" (OuterVolumeSpecName: "kube-api-access-qmrwg") pod "c1b9a905-662c-4e2a-8066-ee8d75e5d45b" (UID: "c1b9a905-662c-4e2a-8066-ee8d75e5d45b"). InnerVolumeSpecName "kube-api-access-qmrwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.747841 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "c1b9a905-662c-4e2a-8066-ee8d75e5d45b" (UID: "c1b9a905-662c-4e2a-8066-ee8d75e5d45b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.748013 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "c1b9a905-662c-4e2a-8066-ee8d75e5d45b" (UID: "c1b9a905-662c-4e2a-8066-ee8d75e5d45b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.748612 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "c1b9a905-662c-4e2a-8066-ee8d75e5d45b" (UID: "c1b9a905-662c-4e2a-8066-ee8d75e5d45b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.748882 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "c1b9a905-662c-4e2a-8066-ee8d75e5d45b" (UID: "c1b9a905-662c-4e2a-8066-ee8d75e5d45b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.749172 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "c1b9a905-662c-4e2a-8066-ee8d75e5d45b" (UID: "c1b9a905-662c-4e2a-8066-ee8d75e5d45b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.749497 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "c1b9a905-662c-4e2a-8066-ee8d75e5d45b" (UID: "c1b9a905-662c-4e2a-8066-ee8d75e5d45b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.751582 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "c1b9a905-662c-4e2a-8066-ee8d75e5d45b" (UID: "c1b9a905-662c-4e2a-8066-ee8d75e5d45b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.842796 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-v4-0-config-system-session\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.842849 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-audit-dir\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.842868 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.842890 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.842910 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-v4-0-config-system-router-certs\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.842932 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-audit-policies\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.842949 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.843084 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.843153 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59g2r\" (UniqueName: \"kubernetes.io/projected/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-kube-api-access-59g2r\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.843196 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.843241 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-v4-0-config-user-template-login\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.843287 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.843313 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-v4-0-config-system-service-ca\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.843337 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-v4-0-config-user-template-error\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.843443 4840 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.843461 4840 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.843474 4840 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.843490 4840 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.843504 4840 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-audit-dir\") on node \"crc\" DevicePath \"\"" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.843518 4840 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.843530 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmrwg\" (UniqueName: \"kubernetes.io/projected/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-kube-api-access-qmrwg\") on node \"crc\" DevicePath \"\"" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.843543 4840 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.843576 4840 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.843588 4840 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-audit-policies\") on node \"crc\" DevicePath \"\"" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.843600 4840 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.843614 4840 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.843626 4840 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.843638 4840 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c1b9a905-662c-4e2a-8066-ee8d75e5d45b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.944971 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.945035 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59g2r\" (UniqueName: \"kubernetes.io/projected/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-kube-api-access-59g2r\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.945059 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.945098 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-v4-0-config-user-template-login\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.945136 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.945157 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-v4-0-config-system-service-ca\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.945174 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-v4-0-config-user-template-error\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.945205 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-v4-0-config-system-session\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.945226 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-audit-dir\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.945250 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.945272 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.945293 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-v4-0-config-system-router-certs\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.945315 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-audit-policies\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.945332 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.945602 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-audit-dir\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.946710 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.946859 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-v4-0-config-system-service-ca\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.947056 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-audit-policies\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.947656 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.948641 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.949209 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.949962 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.950064 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-v4-0-config-system-session\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.951073 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.951161 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-v4-0-config-user-template-login\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.951386 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-v4-0-config-user-template-error\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.951990 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-v4-0-config-system-router-certs\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:25 crc kubenswrapper[4840]: I0930 14:00:25.964503 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59g2r\" (UniqueName: \"kubernetes.io/projected/67e4af69-6f6e-4fdf-90e7-b9c4e1f14354-kube-api-access-59g2r\") pod \"oauth-openshift-7dbc47cf4b-fb6kh\" (UID: \"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354\") " pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:26 crc kubenswrapper[4840]: I0930 14:00:26.024811 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:26 crc kubenswrapper[4840]: I0930 14:00:26.405782 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh"] Sep 30 14:00:26 crc kubenswrapper[4840]: I0930 14:00:26.607710 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" event={"ID":"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354","Type":"ContainerStarted","Data":"bd62ec7deba2239cbec2b66a75c4f8826893aa74ba5b219fc36aef055c7544f8"} Sep 30 14:00:26 crc kubenswrapper[4840]: I0930 14:00:26.609914 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" event={"ID":"c1b9a905-662c-4e2a-8066-ee8d75e5d45b","Type":"ContainerDied","Data":"048ac3b46bf7421fdbedd6180431d06ec5e77076bc085b5d21b543ae66da3ab7"} Sep 30 14:00:26 crc kubenswrapper[4840]: I0930 14:00:26.610019 4840 scope.go:117] "RemoveContainer" containerID="a8aca6442d2827c689d94cb212fdbd982789b44b0113a53ff16a3cb3e0364162" Sep 30 14:00:26 crc kubenswrapper[4840]: I0930 14:00:26.610192 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ncblm" Sep 30 14:00:26 crc kubenswrapper[4840]: I0930 14:00:26.635461 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ncblm"] Sep 30 14:00:26 crc kubenswrapper[4840]: I0930 14:00:26.639369 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ncblm"] Sep 30 14:00:27 crc kubenswrapper[4840]: I0930 14:00:27.621602 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" event={"ID":"67e4af69-6f6e-4fdf-90e7-b9c4e1f14354","Type":"ContainerStarted","Data":"dc1af41248cb4370a35b7b429240ccdd58bc29844995969655368c0af301aca3"} Sep 30 14:00:27 crc kubenswrapper[4840]: I0930 14:00:27.621941 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:27 crc kubenswrapper[4840]: I0930 14:00:27.629317 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" Sep 30 14:00:27 crc kubenswrapper[4840]: I0930 14:00:27.646489 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7dbc47cf4b-fb6kh" podStartSLOduration=27.646468057 podStartE2EDuration="27.646468057s" podCreationTimestamp="2025-09-30 14:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:00:27.645186783 +0000 UTC m=+256.274273226" watchObservedRunningTime="2025-09-30 14:00:27.646468057 +0000 UTC m=+256.275554480" Sep 30 14:00:28 crc kubenswrapper[4840]: I0930 14:00:28.124534 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1b9a905-662c-4e2a-8066-ee8d75e5d45b" path="/var/lib/kubelet/pods/c1b9a905-662c-4e2a-8066-ee8d75e5d45b/volumes" Sep 30 14:00:44 crc kubenswrapper[4840]: I0930 14:00:44.921504 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tc2fx"] Sep 30 14:00:44 crc kubenswrapper[4840]: I0930 14:00:44.922383 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tc2fx" podUID="c97b5b98-9a09-4c3b-9bd0-5a5a8671c479" containerName="registry-server" containerID="cri-o://13fae0bb486eb5bb58d70a75d2ed58249ead39d9548a8d6c09fe2c422ddd815c" gracePeriod=30 Sep 30 14:00:44 crc kubenswrapper[4840]: I0930 14:00:44.929173 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qxvn2"] Sep 30 14:00:44 crc kubenswrapper[4840]: I0930 14:00:44.929496 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-qxvn2" podUID="3c2b4d42-9935-4e56-972f-cdd95d5a0e86" containerName="registry-server" containerID="cri-o://cac37e0372fb79b04eda0d36f6395e9124b5455e61b468361d9bad4e1ea4ba4e" gracePeriod=30 Sep 30 14:00:44 crc kubenswrapper[4840]: I0930 14:00:44.948191 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-snznw"] Sep 30 14:00:44 crc kubenswrapper[4840]: I0930 14:00:44.948429 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-snznw" podUID="32d505eb-6035-488f-b154-2699e8af36ed" containerName="marketplace-operator" containerID="cri-o://f2cea8a3f47b88cffd9635d93bde067d17e1f83cbb3211c684b271f45b73f627" gracePeriod=30 Sep 30 14:00:44 crc kubenswrapper[4840]: I0930 14:00:44.953115 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z895r"] Sep 30 14:00:44 crc kubenswrapper[4840]: I0930 14:00:44.953406 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-z895r" podUID="5d587bdc-9a81-4e7c-9df2-d638d43f5e0d" containerName="registry-server" containerID="cri-o://f18006d0f42332874ec1886fc12d62142fbc2cc6ec63aa08d8041df79e5f8f5c" gracePeriod=30 Sep 30 14:00:44 crc kubenswrapper[4840]: I0930 14:00:44.960705 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5vk9z"] Sep 30 14:00:44 crc kubenswrapper[4840]: I0930 14:00:44.961031 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5vk9z" podUID="3f1c8962-d085-430f-9654-9aa798bc399c" containerName="registry-server" containerID="cri-o://4b0cf2afc6e7946683643afb3dcf6ed4670240ed4be90e19061174d54b826154" gracePeriod=30 Sep 30 14:00:44 crc kubenswrapper[4840]: I0930 14:00:44.964015 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hfdnb"] Sep 30 14:00:44 crc kubenswrapper[4840]: I0930 14:00:44.965445 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-hfdnb" Sep 30 14:00:44 crc kubenswrapper[4840]: I0930 14:00:44.974449 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hfdnb"] Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.104532 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b0c00d91-ae86-48df-84b1-1e1c29b7ff38-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-hfdnb\" (UID: \"b0c00d91-ae86-48df-84b1-1e1c29b7ff38\") " pod="openshift-marketplace/marketplace-operator-79b997595-hfdnb" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.104981 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b0c00d91-ae86-48df-84b1-1e1c29b7ff38-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-hfdnb\" (UID: \"b0c00d91-ae86-48df-84b1-1e1c29b7ff38\") " pod="openshift-marketplace/marketplace-operator-79b997595-hfdnb" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.105075 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cz6ct\" (UniqueName: \"kubernetes.io/projected/b0c00d91-ae86-48df-84b1-1e1c29b7ff38-kube-api-access-cz6ct\") pod \"marketplace-operator-79b997595-hfdnb\" (UID: \"b0c00d91-ae86-48df-84b1-1e1c29b7ff38\") " pod="openshift-marketplace/marketplace-operator-79b997595-hfdnb" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.207310 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b0c00d91-ae86-48df-84b1-1e1c29b7ff38-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-hfdnb\" (UID: \"b0c00d91-ae86-48df-84b1-1e1c29b7ff38\") " pod="openshift-marketplace/marketplace-operator-79b997595-hfdnb" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.207374 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b0c00d91-ae86-48df-84b1-1e1c29b7ff38-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-hfdnb\" (UID: \"b0c00d91-ae86-48df-84b1-1e1c29b7ff38\") " pod="openshift-marketplace/marketplace-operator-79b997595-hfdnb" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.207604 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cz6ct\" (UniqueName: \"kubernetes.io/projected/b0c00d91-ae86-48df-84b1-1e1c29b7ff38-kube-api-access-cz6ct\") pod \"marketplace-operator-79b997595-hfdnb\" (UID: \"b0c00d91-ae86-48df-84b1-1e1c29b7ff38\") " pod="openshift-marketplace/marketplace-operator-79b997595-hfdnb" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.209102 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b0c00d91-ae86-48df-84b1-1e1c29b7ff38-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-hfdnb\" (UID: \"b0c00d91-ae86-48df-84b1-1e1c29b7ff38\") " pod="openshift-marketplace/marketplace-operator-79b997595-hfdnb" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.214614 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b0c00d91-ae86-48df-84b1-1e1c29b7ff38-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-hfdnb\" (UID: \"b0c00d91-ae86-48df-84b1-1e1c29b7ff38\") " pod="openshift-marketplace/marketplace-operator-79b997595-hfdnb" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.227289 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cz6ct\" (UniqueName: \"kubernetes.io/projected/b0c00d91-ae86-48df-84b1-1e1c29b7ff38-kube-api-access-cz6ct\") pod \"marketplace-operator-79b997595-hfdnb\" (UID: \"b0c00d91-ae86-48df-84b1-1e1c29b7ff38\") " pod="openshift-marketplace/marketplace-operator-79b997595-hfdnb" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.286374 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-hfdnb" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.383538 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qxvn2" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.421845 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z895r" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.434029 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tc2fx" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.438838 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-snznw" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.499759 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5vk9z" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.510387 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c2b4d42-9935-4e56-972f-cdd95d5a0e86-catalog-content\") pod \"3c2b4d42-9935-4e56-972f-cdd95d5a0e86\" (UID: \"3c2b4d42-9935-4e56-972f-cdd95d5a0e86\") " Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.510468 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d587bdc-9a81-4e7c-9df2-d638d43f5e0d-catalog-content\") pod \"5d587bdc-9a81-4e7c-9df2-d638d43f5e0d\" (UID: \"5d587bdc-9a81-4e7c-9df2-d638d43f5e0d\") " Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.510510 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6t49m\" (UniqueName: \"kubernetes.io/projected/5d587bdc-9a81-4e7c-9df2-d638d43f5e0d-kube-api-access-6t49m\") pod \"5d587bdc-9a81-4e7c-9df2-d638d43f5e0d\" (UID: \"5d587bdc-9a81-4e7c-9df2-d638d43f5e0d\") " Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.510617 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d587bdc-9a81-4e7c-9df2-d638d43f5e0d-utilities\") pod \"5d587bdc-9a81-4e7c-9df2-d638d43f5e0d\" (UID: \"5d587bdc-9a81-4e7c-9df2-d638d43f5e0d\") " Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.510692 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c97b5b98-9a09-4c3b-9bd0-5a5a8671c479-catalog-content\") pod \"c97b5b98-9a09-4c3b-9bd0-5a5a8671c479\" (UID: \"c97b5b98-9a09-4c3b-9bd0-5a5a8671c479\") " Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.510728 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/32d505eb-6035-488f-b154-2699e8af36ed-marketplace-operator-metrics\") pod \"32d505eb-6035-488f-b154-2699e8af36ed\" (UID: \"32d505eb-6035-488f-b154-2699e8af36ed\") " Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.510775 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxk8d\" (UniqueName: \"kubernetes.io/projected/c97b5b98-9a09-4c3b-9bd0-5a5a8671c479-kube-api-access-nxk8d\") pod \"c97b5b98-9a09-4c3b-9bd0-5a5a8671c479\" (UID: \"c97b5b98-9a09-4c3b-9bd0-5a5a8671c479\") " Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.510811 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qkk5g\" (UniqueName: \"kubernetes.io/projected/32d505eb-6035-488f-b154-2699e8af36ed-kube-api-access-qkk5g\") pod \"32d505eb-6035-488f-b154-2699e8af36ed\" (UID: \"32d505eb-6035-488f-b154-2699e8af36ed\") " Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.510861 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c2b4d42-9935-4e56-972f-cdd95d5a0e86-utilities\") pod \"3c2b4d42-9935-4e56-972f-cdd95d5a0e86\" (UID: \"3c2b4d42-9935-4e56-972f-cdd95d5a0e86\") " Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.510886 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/32d505eb-6035-488f-b154-2699e8af36ed-marketplace-trusted-ca\") pod \"32d505eb-6035-488f-b154-2699e8af36ed\" (UID: \"32d505eb-6035-488f-b154-2699e8af36ed\") " Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.510921 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c97b5b98-9a09-4c3b-9bd0-5a5a8671c479-utilities\") pod \"c97b5b98-9a09-4c3b-9bd0-5a5a8671c479\" (UID: \"c97b5b98-9a09-4c3b-9bd0-5a5a8671c479\") " Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.510960 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bt794\" (UniqueName: \"kubernetes.io/projected/3c2b4d42-9935-4e56-972f-cdd95d5a0e86-kube-api-access-bt794\") pod \"3c2b4d42-9935-4e56-972f-cdd95d5a0e86\" (UID: \"3c2b4d42-9935-4e56-972f-cdd95d5a0e86\") " Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.512816 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c2b4d42-9935-4e56-972f-cdd95d5a0e86-utilities" (OuterVolumeSpecName: "utilities") pod "3c2b4d42-9935-4e56-972f-cdd95d5a0e86" (UID: "3c2b4d42-9935-4e56-972f-cdd95d5a0e86"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.514033 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32d505eb-6035-488f-b154-2699e8af36ed-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "32d505eb-6035-488f-b154-2699e8af36ed" (UID: "32d505eb-6035-488f-b154-2699e8af36ed"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.515335 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d587bdc-9a81-4e7c-9df2-d638d43f5e0d-utilities" (OuterVolumeSpecName: "utilities") pod "5d587bdc-9a81-4e7c-9df2-d638d43f5e0d" (UID: "5d587bdc-9a81-4e7c-9df2-d638d43f5e0d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.521882 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c2b4d42-9935-4e56-972f-cdd95d5a0e86-kube-api-access-bt794" (OuterVolumeSpecName: "kube-api-access-bt794") pod "3c2b4d42-9935-4e56-972f-cdd95d5a0e86" (UID: "3c2b4d42-9935-4e56-972f-cdd95d5a0e86"). InnerVolumeSpecName "kube-api-access-bt794". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.522281 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32d505eb-6035-488f-b154-2699e8af36ed-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "32d505eb-6035-488f-b154-2699e8af36ed" (UID: "32d505eb-6035-488f-b154-2699e8af36ed"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.522343 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c97b5b98-9a09-4c3b-9bd0-5a5a8671c479-kube-api-access-nxk8d" (OuterVolumeSpecName: "kube-api-access-nxk8d") pod "c97b5b98-9a09-4c3b-9bd0-5a5a8671c479" (UID: "c97b5b98-9a09-4c3b-9bd0-5a5a8671c479"). InnerVolumeSpecName "kube-api-access-nxk8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.526062 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32d505eb-6035-488f-b154-2699e8af36ed-kube-api-access-qkk5g" (OuterVolumeSpecName: "kube-api-access-qkk5g") pod "32d505eb-6035-488f-b154-2699e8af36ed" (UID: "32d505eb-6035-488f-b154-2699e8af36ed"). InnerVolumeSpecName "kube-api-access-qkk5g". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.530798 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d587bdc-9a81-4e7c-9df2-d638d43f5e0d-kube-api-access-6t49m" (OuterVolumeSpecName: "kube-api-access-6t49m") pod "5d587bdc-9a81-4e7c-9df2-d638d43f5e0d" (UID: "5d587bdc-9a81-4e7c-9df2-d638d43f5e0d"). InnerVolumeSpecName "kube-api-access-6t49m". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.531899 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c97b5b98-9a09-4c3b-9bd0-5a5a8671c479-utilities" (OuterVolumeSpecName: "utilities") pod "c97b5b98-9a09-4c3b-9bd0-5a5a8671c479" (UID: "c97b5b98-9a09-4c3b-9bd0-5a5a8671c479"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.540934 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d587bdc-9a81-4e7c-9df2-d638d43f5e0d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5d587bdc-9a81-4e7c-9df2-d638d43f5e0d" (UID: "5d587bdc-9a81-4e7c-9df2-d638d43f5e0d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.578885 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hfdnb"] Sep 30 14:00:45 crc kubenswrapper[4840]: W0930 14:00:45.582407 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb0c00d91_ae86_48df_84b1_1e1c29b7ff38.slice/crio-40a7e3b23da9737076a3daf90ff4fd65d318520042189ca2ece1bc5eb7705d2d WatchSource:0}: Error finding container 40a7e3b23da9737076a3daf90ff4fd65d318520042189ca2ece1bc5eb7705d2d: Status 404 returned error can't find the container with id 40a7e3b23da9737076a3daf90ff4fd65d318520042189ca2ece1bc5eb7705d2d Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.591259 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c2b4d42-9935-4e56-972f-cdd95d5a0e86-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3c2b4d42-9935-4e56-972f-cdd95d5a0e86" (UID: "3c2b4d42-9935-4e56-972f-cdd95d5a0e86"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.605323 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c97b5b98-9a09-4c3b-9bd0-5a5a8671c479-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c97b5b98-9a09-4c3b-9bd0-5a5a8671c479" (UID: "c97b5b98-9a09-4c3b-9bd0-5a5a8671c479"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.612236 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f1c8962-d085-430f-9654-9aa798bc399c-catalog-content\") pod \"3f1c8962-d085-430f-9654-9aa798bc399c\" (UID: \"3f1c8962-d085-430f-9654-9aa798bc399c\") " Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.612285 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f1c8962-d085-430f-9654-9aa798bc399c-utilities\") pod \"3f1c8962-d085-430f-9654-9aa798bc399c\" (UID: \"3f1c8962-d085-430f-9654-9aa798bc399c\") " Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.612370 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x64f5\" (UniqueName: \"kubernetes.io/projected/3f1c8962-d085-430f-9654-9aa798bc399c-kube-api-access-x64f5\") pod \"3f1c8962-d085-430f-9654-9aa798bc399c\" (UID: \"3f1c8962-d085-430f-9654-9aa798bc399c\") " Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.612629 4840 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c2b4d42-9935-4e56-972f-cdd95d5a0e86-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.612650 4840 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d587bdc-9a81-4e7c-9df2-d638d43f5e0d-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.612662 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6t49m\" (UniqueName: \"kubernetes.io/projected/5d587bdc-9a81-4e7c-9df2-d638d43f5e0d-kube-api-access-6t49m\") on node \"crc\" DevicePath \"\"" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.612676 4840 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d587bdc-9a81-4e7c-9df2-d638d43f5e0d-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.612687 4840 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c97b5b98-9a09-4c3b-9bd0-5a5a8671c479-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.612699 4840 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/32d505eb-6035-488f-b154-2699e8af36ed-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.612713 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxk8d\" (UniqueName: \"kubernetes.io/projected/c97b5b98-9a09-4c3b-9bd0-5a5a8671c479-kube-api-access-nxk8d\") on node \"crc\" DevicePath \"\"" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.612725 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qkk5g\" (UniqueName: \"kubernetes.io/projected/32d505eb-6035-488f-b154-2699e8af36ed-kube-api-access-qkk5g\") on node \"crc\" DevicePath \"\"" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.612827 4840 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c2b4d42-9935-4e56-972f-cdd95d5a0e86-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.612837 4840 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/32d505eb-6035-488f-b154-2699e8af36ed-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.612845 4840 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c97b5b98-9a09-4c3b-9bd0-5a5a8671c479-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.612853 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bt794\" (UniqueName: \"kubernetes.io/projected/3c2b4d42-9935-4e56-972f-cdd95d5a0e86-kube-api-access-bt794\") on node \"crc\" DevicePath \"\"" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.614280 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f1c8962-d085-430f-9654-9aa798bc399c-utilities" (OuterVolumeSpecName: "utilities") pod "3f1c8962-d085-430f-9654-9aa798bc399c" (UID: "3f1c8962-d085-430f-9654-9aa798bc399c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.614996 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f1c8962-d085-430f-9654-9aa798bc399c-kube-api-access-x64f5" (OuterVolumeSpecName: "kube-api-access-x64f5") pod "3f1c8962-d085-430f-9654-9aa798bc399c" (UID: "3f1c8962-d085-430f-9654-9aa798bc399c"). InnerVolumeSpecName "kube-api-access-x64f5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.704138 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f1c8962-d085-430f-9654-9aa798bc399c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3f1c8962-d085-430f-9654-9aa798bc399c" (UID: "3f1c8962-d085-430f-9654-9aa798bc399c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.714063 4840 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f1c8962-d085-430f-9654-9aa798bc399c-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.714101 4840 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f1c8962-d085-430f-9654-9aa798bc399c-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.714112 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x64f5\" (UniqueName: \"kubernetes.io/projected/3f1c8962-d085-430f-9654-9aa798bc399c-kube-api-access-x64f5\") on node \"crc\" DevicePath \"\"" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.718717 4840 generic.go:334] "Generic (PLEG): container finished" podID="3f1c8962-d085-430f-9654-9aa798bc399c" containerID="4b0cf2afc6e7946683643afb3dcf6ed4670240ed4be90e19061174d54b826154" exitCode=0 Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.718773 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5vk9z" event={"ID":"3f1c8962-d085-430f-9654-9aa798bc399c","Type":"ContainerDied","Data":"4b0cf2afc6e7946683643afb3dcf6ed4670240ed4be90e19061174d54b826154"} Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.718799 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5vk9z" event={"ID":"3f1c8962-d085-430f-9654-9aa798bc399c","Type":"ContainerDied","Data":"14c02f20e6ca97546431aa1e107c57bb3e48f643e7cac9f56ae269bd642a077f"} Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.718814 4840 scope.go:117] "RemoveContainer" containerID="4b0cf2afc6e7946683643afb3dcf6ed4670240ed4be90e19061174d54b826154" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.718904 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5vk9z" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.723007 4840 generic.go:334] "Generic (PLEG): container finished" podID="32d505eb-6035-488f-b154-2699e8af36ed" containerID="f2cea8a3f47b88cffd9635d93bde067d17e1f83cbb3211c684b271f45b73f627" exitCode=0 Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.723066 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-snznw" event={"ID":"32d505eb-6035-488f-b154-2699e8af36ed","Type":"ContainerDied","Data":"f2cea8a3f47b88cffd9635d93bde067d17e1f83cbb3211c684b271f45b73f627"} Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.723093 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-snznw" event={"ID":"32d505eb-6035-488f-b154-2699e8af36ed","Type":"ContainerDied","Data":"38a1b3e8457273c6176cf0b8c5674d0517c65eb9f14330018e2ceb47ddc97a14"} Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.723158 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-snznw" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.726383 4840 generic.go:334] "Generic (PLEG): container finished" podID="3c2b4d42-9935-4e56-972f-cdd95d5a0e86" containerID="cac37e0372fb79b04eda0d36f6395e9124b5455e61b468361d9bad4e1ea4ba4e" exitCode=0 Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.726450 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qxvn2" event={"ID":"3c2b4d42-9935-4e56-972f-cdd95d5a0e86","Type":"ContainerDied","Data":"cac37e0372fb79b04eda0d36f6395e9124b5455e61b468361d9bad4e1ea4ba4e"} Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.726466 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qxvn2" event={"ID":"3c2b4d42-9935-4e56-972f-cdd95d5a0e86","Type":"ContainerDied","Data":"a4b2eff50a8ea900a24e4a66c753298952e6b102215140619f8a8f8e6e7b49b9"} Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.726531 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qxvn2" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.729479 4840 generic.go:334] "Generic (PLEG): container finished" podID="c97b5b98-9a09-4c3b-9bd0-5a5a8671c479" containerID="13fae0bb486eb5bb58d70a75d2ed58249ead39d9548a8d6c09fe2c422ddd815c" exitCode=0 Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.729666 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tc2fx" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.729665 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tc2fx" event={"ID":"c97b5b98-9a09-4c3b-9bd0-5a5a8671c479","Type":"ContainerDied","Data":"13fae0bb486eb5bb58d70a75d2ed58249ead39d9548a8d6c09fe2c422ddd815c"} Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.729878 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tc2fx" event={"ID":"c97b5b98-9a09-4c3b-9bd0-5a5a8671c479","Type":"ContainerDied","Data":"71b299bc9ebe1e1e182a3e5e3f8a058663b44eeb9aac7e2442d4eb684c0025b2"} Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.733759 4840 generic.go:334] "Generic (PLEG): container finished" podID="5d587bdc-9a81-4e7c-9df2-d638d43f5e0d" containerID="f18006d0f42332874ec1886fc12d62142fbc2cc6ec63aa08d8041df79e5f8f5c" exitCode=0 Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.733819 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z895r" event={"ID":"5d587bdc-9a81-4e7c-9df2-d638d43f5e0d","Type":"ContainerDied","Data":"f18006d0f42332874ec1886fc12d62142fbc2cc6ec63aa08d8041df79e5f8f5c"} Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.733847 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z895r" event={"ID":"5d587bdc-9a81-4e7c-9df2-d638d43f5e0d","Type":"ContainerDied","Data":"27587a514f130738ead13f9bc8547449515e24e6965c0f9263e18cdfe4b1f107"} Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.733904 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z895r" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.738608 4840 scope.go:117] "RemoveContainer" containerID="645ed392836c63be996642ef20a1a081037ffb00d13e2eae9c24b844c0c9f701" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.738763 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-hfdnb" event={"ID":"b0c00d91-ae86-48df-84b1-1e1c29b7ff38","Type":"ContainerStarted","Data":"40a7e3b23da9737076a3daf90ff4fd65d318520042189ca2ece1bc5eb7705d2d"} Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.768190 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5vk9z"] Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.771753 4840 scope.go:117] "RemoveContainer" containerID="761f0cb380334134a3c3d0c08581b8420e0880acb55b8526f5a42528399a0f18" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.779203 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5vk9z"] Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.782520 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-snznw"] Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.787418 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-snznw"] Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.793782 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z895r"] Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.797053 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-z895r"] Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.799502 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qxvn2"] Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.801939 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-qxvn2"] Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.812216 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tc2fx"] Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.814576 4840 scope.go:117] "RemoveContainer" containerID="4b0cf2afc6e7946683643afb3dcf6ed4670240ed4be90e19061174d54b826154" Sep 30 14:00:45 crc kubenswrapper[4840]: E0930 14:00:45.815419 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b0cf2afc6e7946683643afb3dcf6ed4670240ed4be90e19061174d54b826154\": container with ID starting with 4b0cf2afc6e7946683643afb3dcf6ed4670240ed4be90e19061174d54b826154 not found: ID does not exist" containerID="4b0cf2afc6e7946683643afb3dcf6ed4670240ed4be90e19061174d54b826154" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.815464 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b0cf2afc6e7946683643afb3dcf6ed4670240ed4be90e19061174d54b826154"} err="failed to get container status \"4b0cf2afc6e7946683643afb3dcf6ed4670240ed4be90e19061174d54b826154\": rpc error: code = NotFound desc = could not find container \"4b0cf2afc6e7946683643afb3dcf6ed4670240ed4be90e19061174d54b826154\": container with ID starting with 4b0cf2afc6e7946683643afb3dcf6ed4670240ed4be90e19061174d54b826154 not found: ID does not exist" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.815495 4840 scope.go:117] "RemoveContainer" containerID="645ed392836c63be996642ef20a1a081037ffb00d13e2eae9c24b844c0c9f701" Sep 30 14:00:45 crc kubenswrapper[4840]: E0930 14:00:45.815833 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"645ed392836c63be996642ef20a1a081037ffb00d13e2eae9c24b844c0c9f701\": container with ID starting with 645ed392836c63be996642ef20a1a081037ffb00d13e2eae9c24b844c0c9f701 not found: ID does not exist" containerID="645ed392836c63be996642ef20a1a081037ffb00d13e2eae9c24b844c0c9f701" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.815864 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"645ed392836c63be996642ef20a1a081037ffb00d13e2eae9c24b844c0c9f701"} err="failed to get container status \"645ed392836c63be996642ef20a1a081037ffb00d13e2eae9c24b844c0c9f701\": rpc error: code = NotFound desc = could not find container \"645ed392836c63be996642ef20a1a081037ffb00d13e2eae9c24b844c0c9f701\": container with ID starting with 645ed392836c63be996642ef20a1a081037ffb00d13e2eae9c24b844c0c9f701 not found: ID does not exist" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.815896 4840 scope.go:117] "RemoveContainer" containerID="761f0cb380334134a3c3d0c08581b8420e0880acb55b8526f5a42528399a0f18" Sep 30 14:00:45 crc kubenswrapper[4840]: E0930 14:00:45.816320 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"761f0cb380334134a3c3d0c08581b8420e0880acb55b8526f5a42528399a0f18\": container with ID starting with 761f0cb380334134a3c3d0c08581b8420e0880acb55b8526f5a42528399a0f18 not found: ID does not exist" containerID="761f0cb380334134a3c3d0c08581b8420e0880acb55b8526f5a42528399a0f18" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.816340 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tc2fx"] Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.816361 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"761f0cb380334134a3c3d0c08581b8420e0880acb55b8526f5a42528399a0f18"} err="failed to get container status \"761f0cb380334134a3c3d0c08581b8420e0880acb55b8526f5a42528399a0f18\": rpc error: code = NotFound desc = could not find container \"761f0cb380334134a3c3d0c08581b8420e0880acb55b8526f5a42528399a0f18\": container with ID starting with 761f0cb380334134a3c3d0c08581b8420e0880acb55b8526f5a42528399a0f18 not found: ID does not exist" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.816387 4840 scope.go:117] "RemoveContainer" containerID="f2cea8a3f47b88cffd9635d93bde067d17e1f83cbb3211c684b271f45b73f627" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.832530 4840 scope.go:117] "RemoveContainer" containerID="f2cea8a3f47b88cffd9635d93bde067d17e1f83cbb3211c684b271f45b73f627" Sep 30 14:00:45 crc kubenswrapper[4840]: E0930 14:00:45.833016 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2cea8a3f47b88cffd9635d93bde067d17e1f83cbb3211c684b271f45b73f627\": container with ID starting with f2cea8a3f47b88cffd9635d93bde067d17e1f83cbb3211c684b271f45b73f627 not found: ID does not exist" containerID="f2cea8a3f47b88cffd9635d93bde067d17e1f83cbb3211c684b271f45b73f627" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.833063 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2cea8a3f47b88cffd9635d93bde067d17e1f83cbb3211c684b271f45b73f627"} err="failed to get container status \"f2cea8a3f47b88cffd9635d93bde067d17e1f83cbb3211c684b271f45b73f627\": rpc error: code = NotFound desc = could not find container \"f2cea8a3f47b88cffd9635d93bde067d17e1f83cbb3211c684b271f45b73f627\": container with ID starting with f2cea8a3f47b88cffd9635d93bde067d17e1f83cbb3211c684b271f45b73f627 not found: ID does not exist" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.833094 4840 scope.go:117] "RemoveContainer" containerID="cac37e0372fb79b04eda0d36f6395e9124b5455e61b468361d9bad4e1ea4ba4e" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.848305 4840 scope.go:117] "RemoveContainer" containerID="8a4822dab2e81685d61813fc0e1f6505f9349dee991207ba566c3443e8ea414f" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.868423 4840 scope.go:117] "RemoveContainer" containerID="d99911001fbf445dedee82d6b2eeabfcc62f08e815da035535039377331df036" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.889804 4840 scope.go:117] "RemoveContainer" containerID="cac37e0372fb79b04eda0d36f6395e9124b5455e61b468361d9bad4e1ea4ba4e" Sep 30 14:00:45 crc kubenswrapper[4840]: E0930 14:00:45.890730 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cac37e0372fb79b04eda0d36f6395e9124b5455e61b468361d9bad4e1ea4ba4e\": container with ID starting with cac37e0372fb79b04eda0d36f6395e9124b5455e61b468361d9bad4e1ea4ba4e not found: ID does not exist" containerID="cac37e0372fb79b04eda0d36f6395e9124b5455e61b468361d9bad4e1ea4ba4e" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.890757 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cac37e0372fb79b04eda0d36f6395e9124b5455e61b468361d9bad4e1ea4ba4e"} err="failed to get container status \"cac37e0372fb79b04eda0d36f6395e9124b5455e61b468361d9bad4e1ea4ba4e\": rpc error: code = NotFound desc = could not find container \"cac37e0372fb79b04eda0d36f6395e9124b5455e61b468361d9bad4e1ea4ba4e\": container with ID starting with cac37e0372fb79b04eda0d36f6395e9124b5455e61b468361d9bad4e1ea4ba4e not found: ID does not exist" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.890780 4840 scope.go:117] "RemoveContainer" containerID="8a4822dab2e81685d61813fc0e1f6505f9349dee991207ba566c3443e8ea414f" Sep 30 14:00:45 crc kubenswrapper[4840]: E0930 14:00:45.891150 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a4822dab2e81685d61813fc0e1f6505f9349dee991207ba566c3443e8ea414f\": container with ID starting with 8a4822dab2e81685d61813fc0e1f6505f9349dee991207ba566c3443e8ea414f not found: ID does not exist" containerID="8a4822dab2e81685d61813fc0e1f6505f9349dee991207ba566c3443e8ea414f" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.891169 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a4822dab2e81685d61813fc0e1f6505f9349dee991207ba566c3443e8ea414f"} err="failed to get container status \"8a4822dab2e81685d61813fc0e1f6505f9349dee991207ba566c3443e8ea414f\": rpc error: code = NotFound desc = could not find container \"8a4822dab2e81685d61813fc0e1f6505f9349dee991207ba566c3443e8ea414f\": container with ID starting with 8a4822dab2e81685d61813fc0e1f6505f9349dee991207ba566c3443e8ea414f not found: ID does not exist" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.891182 4840 scope.go:117] "RemoveContainer" containerID="d99911001fbf445dedee82d6b2eeabfcc62f08e815da035535039377331df036" Sep 30 14:00:45 crc kubenswrapper[4840]: E0930 14:00:45.891588 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d99911001fbf445dedee82d6b2eeabfcc62f08e815da035535039377331df036\": container with ID starting with d99911001fbf445dedee82d6b2eeabfcc62f08e815da035535039377331df036 not found: ID does not exist" containerID="d99911001fbf445dedee82d6b2eeabfcc62f08e815da035535039377331df036" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.891616 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d99911001fbf445dedee82d6b2eeabfcc62f08e815da035535039377331df036"} err="failed to get container status \"d99911001fbf445dedee82d6b2eeabfcc62f08e815da035535039377331df036\": rpc error: code = NotFound desc = could not find container \"d99911001fbf445dedee82d6b2eeabfcc62f08e815da035535039377331df036\": container with ID starting with d99911001fbf445dedee82d6b2eeabfcc62f08e815da035535039377331df036 not found: ID does not exist" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.891635 4840 scope.go:117] "RemoveContainer" containerID="13fae0bb486eb5bb58d70a75d2ed58249ead39d9548a8d6c09fe2c422ddd815c" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.904970 4840 scope.go:117] "RemoveContainer" containerID="8d3a35c625e0dee7fbe9ee695b32c4d5c35612188e2e65fd036fcb781d3abdaf" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.919205 4840 scope.go:117] "RemoveContainer" containerID="0ed1236664362f6f4988de894b3dbaf9ce29ab242e2a03f872b438a16b279af7" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.930971 4840 scope.go:117] "RemoveContainer" containerID="13fae0bb486eb5bb58d70a75d2ed58249ead39d9548a8d6c09fe2c422ddd815c" Sep 30 14:00:45 crc kubenswrapper[4840]: E0930 14:00:45.931427 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13fae0bb486eb5bb58d70a75d2ed58249ead39d9548a8d6c09fe2c422ddd815c\": container with ID starting with 13fae0bb486eb5bb58d70a75d2ed58249ead39d9548a8d6c09fe2c422ddd815c not found: ID does not exist" containerID="13fae0bb486eb5bb58d70a75d2ed58249ead39d9548a8d6c09fe2c422ddd815c" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.931480 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13fae0bb486eb5bb58d70a75d2ed58249ead39d9548a8d6c09fe2c422ddd815c"} err="failed to get container status \"13fae0bb486eb5bb58d70a75d2ed58249ead39d9548a8d6c09fe2c422ddd815c\": rpc error: code = NotFound desc = could not find container \"13fae0bb486eb5bb58d70a75d2ed58249ead39d9548a8d6c09fe2c422ddd815c\": container with ID starting with 13fae0bb486eb5bb58d70a75d2ed58249ead39d9548a8d6c09fe2c422ddd815c not found: ID does not exist" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.931513 4840 scope.go:117] "RemoveContainer" containerID="8d3a35c625e0dee7fbe9ee695b32c4d5c35612188e2e65fd036fcb781d3abdaf" Sep 30 14:00:45 crc kubenswrapper[4840]: E0930 14:00:45.931978 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d3a35c625e0dee7fbe9ee695b32c4d5c35612188e2e65fd036fcb781d3abdaf\": container with ID starting with 8d3a35c625e0dee7fbe9ee695b32c4d5c35612188e2e65fd036fcb781d3abdaf not found: ID does not exist" containerID="8d3a35c625e0dee7fbe9ee695b32c4d5c35612188e2e65fd036fcb781d3abdaf" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.932022 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d3a35c625e0dee7fbe9ee695b32c4d5c35612188e2e65fd036fcb781d3abdaf"} err="failed to get container status \"8d3a35c625e0dee7fbe9ee695b32c4d5c35612188e2e65fd036fcb781d3abdaf\": rpc error: code = NotFound desc = could not find container \"8d3a35c625e0dee7fbe9ee695b32c4d5c35612188e2e65fd036fcb781d3abdaf\": container with ID starting with 8d3a35c625e0dee7fbe9ee695b32c4d5c35612188e2e65fd036fcb781d3abdaf not found: ID does not exist" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.932063 4840 scope.go:117] "RemoveContainer" containerID="0ed1236664362f6f4988de894b3dbaf9ce29ab242e2a03f872b438a16b279af7" Sep 30 14:00:45 crc kubenswrapper[4840]: E0930 14:00:45.932323 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ed1236664362f6f4988de894b3dbaf9ce29ab242e2a03f872b438a16b279af7\": container with ID starting with 0ed1236664362f6f4988de894b3dbaf9ce29ab242e2a03f872b438a16b279af7 not found: ID does not exist" containerID="0ed1236664362f6f4988de894b3dbaf9ce29ab242e2a03f872b438a16b279af7" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.932352 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ed1236664362f6f4988de894b3dbaf9ce29ab242e2a03f872b438a16b279af7"} err="failed to get container status \"0ed1236664362f6f4988de894b3dbaf9ce29ab242e2a03f872b438a16b279af7\": rpc error: code = NotFound desc = could not find container \"0ed1236664362f6f4988de894b3dbaf9ce29ab242e2a03f872b438a16b279af7\": container with ID starting with 0ed1236664362f6f4988de894b3dbaf9ce29ab242e2a03f872b438a16b279af7 not found: ID does not exist" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.932370 4840 scope.go:117] "RemoveContainer" containerID="f18006d0f42332874ec1886fc12d62142fbc2cc6ec63aa08d8041df79e5f8f5c" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.946685 4840 scope.go:117] "RemoveContainer" containerID="11b5a464fb1cdc01fdad7819b80af6a830c396a49fe2a970dd8011998936cb20" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.965011 4840 scope.go:117] "RemoveContainer" containerID="b3db70d2b71ecaa2fbbdb1ffe2a875f5647437b38eae474dd5ffb08cc1b54a6d" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.980808 4840 scope.go:117] "RemoveContainer" containerID="f18006d0f42332874ec1886fc12d62142fbc2cc6ec63aa08d8041df79e5f8f5c" Sep 30 14:00:45 crc kubenswrapper[4840]: E0930 14:00:45.981273 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f18006d0f42332874ec1886fc12d62142fbc2cc6ec63aa08d8041df79e5f8f5c\": container with ID starting with f18006d0f42332874ec1886fc12d62142fbc2cc6ec63aa08d8041df79e5f8f5c not found: ID does not exist" containerID="f18006d0f42332874ec1886fc12d62142fbc2cc6ec63aa08d8041df79e5f8f5c" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.981325 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f18006d0f42332874ec1886fc12d62142fbc2cc6ec63aa08d8041df79e5f8f5c"} err="failed to get container status \"f18006d0f42332874ec1886fc12d62142fbc2cc6ec63aa08d8041df79e5f8f5c\": rpc error: code = NotFound desc = could not find container \"f18006d0f42332874ec1886fc12d62142fbc2cc6ec63aa08d8041df79e5f8f5c\": container with ID starting with f18006d0f42332874ec1886fc12d62142fbc2cc6ec63aa08d8041df79e5f8f5c not found: ID does not exist" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.981363 4840 scope.go:117] "RemoveContainer" containerID="11b5a464fb1cdc01fdad7819b80af6a830c396a49fe2a970dd8011998936cb20" Sep 30 14:00:45 crc kubenswrapper[4840]: E0930 14:00:45.981946 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11b5a464fb1cdc01fdad7819b80af6a830c396a49fe2a970dd8011998936cb20\": container with ID starting with 11b5a464fb1cdc01fdad7819b80af6a830c396a49fe2a970dd8011998936cb20 not found: ID does not exist" containerID="11b5a464fb1cdc01fdad7819b80af6a830c396a49fe2a970dd8011998936cb20" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.981982 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11b5a464fb1cdc01fdad7819b80af6a830c396a49fe2a970dd8011998936cb20"} err="failed to get container status \"11b5a464fb1cdc01fdad7819b80af6a830c396a49fe2a970dd8011998936cb20\": rpc error: code = NotFound desc = could not find container \"11b5a464fb1cdc01fdad7819b80af6a830c396a49fe2a970dd8011998936cb20\": container with ID starting with 11b5a464fb1cdc01fdad7819b80af6a830c396a49fe2a970dd8011998936cb20 not found: ID does not exist" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.982012 4840 scope.go:117] "RemoveContainer" containerID="b3db70d2b71ecaa2fbbdb1ffe2a875f5647437b38eae474dd5ffb08cc1b54a6d" Sep 30 14:00:45 crc kubenswrapper[4840]: E0930 14:00:45.982346 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3db70d2b71ecaa2fbbdb1ffe2a875f5647437b38eae474dd5ffb08cc1b54a6d\": container with ID starting with b3db70d2b71ecaa2fbbdb1ffe2a875f5647437b38eae474dd5ffb08cc1b54a6d not found: ID does not exist" containerID="b3db70d2b71ecaa2fbbdb1ffe2a875f5647437b38eae474dd5ffb08cc1b54a6d" Sep 30 14:00:45 crc kubenswrapper[4840]: I0930 14:00:45.982369 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3db70d2b71ecaa2fbbdb1ffe2a875f5647437b38eae474dd5ffb08cc1b54a6d"} err="failed to get container status \"b3db70d2b71ecaa2fbbdb1ffe2a875f5647437b38eae474dd5ffb08cc1b54a6d\": rpc error: code = NotFound desc = could not find container \"b3db70d2b71ecaa2fbbdb1ffe2a875f5647437b38eae474dd5ffb08cc1b54a6d\": container with ID starting with b3db70d2b71ecaa2fbbdb1ffe2a875f5647437b38eae474dd5ffb08cc1b54a6d not found: ID does not exist" Sep 30 14:00:46 crc kubenswrapper[4840]: I0930 14:00:46.123676 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32d505eb-6035-488f-b154-2699e8af36ed" path="/var/lib/kubelet/pods/32d505eb-6035-488f-b154-2699e8af36ed/volumes" Sep 30 14:00:46 crc kubenswrapper[4840]: I0930 14:00:46.124223 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c2b4d42-9935-4e56-972f-cdd95d5a0e86" path="/var/lib/kubelet/pods/3c2b4d42-9935-4e56-972f-cdd95d5a0e86/volumes" Sep 30 14:00:46 crc kubenswrapper[4840]: I0930 14:00:46.124827 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f1c8962-d085-430f-9654-9aa798bc399c" path="/var/lib/kubelet/pods/3f1c8962-d085-430f-9654-9aa798bc399c/volumes" Sep 30 14:00:46 crc kubenswrapper[4840]: I0930 14:00:46.125871 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d587bdc-9a81-4e7c-9df2-d638d43f5e0d" path="/var/lib/kubelet/pods/5d587bdc-9a81-4e7c-9df2-d638d43f5e0d/volumes" Sep 30 14:00:46 crc kubenswrapper[4840]: I0930 14:00:46.126657 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c97b5b98-9a09-4c3b-9bd0-5a5a8671c479" path="/var/lib/kubelet/pods/c97b5b98-9a09-4c3b-9bd0-5a5a8671c479/volumes" Sep 30 14:00:46 crc kubenswrapper[4840]: I0930 14:00:46.751852 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-hfdnb" event={"ID":"b0c00d91-ae86-48df-84b1-1e1c29b7ff38","Type":"ContainerStarted","Data":"b052dc0822c806f81223726a0f50db64688efdc401ee3a5906777ef1b9dab95a"} Sep 30 14:00:46 crc kubenswrapper[4840]: I0930 14:00:46.752147 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-hfdnb" Sep 30 14:00:46 crc kubenswrapper[4840]: I0930 14:00:46.757815 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-hfdnb" Sep 30 14:00:46 crc kubenswrapper[4840]: I0930 14:00:46.768040 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-hfdnb" podStartSLOduration=2.76802002 podStartE2EDuration="2.76802002s" podCreationTimestamp="2025-09-30 14:00:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:00:46.765367109 +0000 UTC m=+275.394453532" watchObservedRunningTime="2025-09-30 14:00:46.76802002 +0000 UTC m=+275.397106443" Sep 30 14:00:47 crc kubenswrapper[4840]: I0930 14:00:47.367506 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-f85rt"] Sep 30 14:00:47 crc kubenswrapper[4840]: E0930 14:00:47.367744 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c97b5b98-9a09-4c3b-9bd0-5a5a8671c479" containerName="extract-content" Sep 30 14:00:47 crc kubenswrapper[4840]: I0930 14:00:47.367761 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="c97b5b98-9a09-4c3b-9bd0-5a5a8671c479" containerName="extract-content" Sep 30 14:00:47 crc kubenswrapper[4840]: E0930 14:00:47.367778 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f1c8962-d085-430f-9654-9aa798bc399c" containerName="extract-utilities" Sep 30 14:00:47 crc kubenswrapper[4840]: I0930 14:00:47.367787 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f1c8962-d085-430f-9654-9aa798bc399c" containerName="extract-utilities" Sep 30 14:00:47 crc kubenswrapper[4840]: E0930 14:00:47.367800 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f1c8962-d085-430f-9654-9aa798bc399c" containerName="extract-content" Sep 30 14:00:47 crc kubenswrapper[4840]: I0930 14:00:47.367807 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f1c8962-d085-430f-9654-9aa798bc399c" containerName="extract-content" Sep 30 14:00:47 crc kubenswrapper[4840]: E0930 14:00:47.367818 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c97b5b98-9a09-4c3b-9bd0-5a5a8671c479" containerName="extract-utilities" Sep 30 14:00:47 crc kubenswrapper[4840]: I0930 14:00:47.367826 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="c97b5b98-9a09-4c3b-9bd0-5a5a8671c479" containerName="extract-utilities" Sep 30 14:00:47 crc kubenswrapper[4840]: E0930 14:00:47.367841 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c2b4d42-9935-4e56-972f-cdd95d5a0e86" containerName="extract-utilities" Sep 30 14:00:47 crc kubenswrapper[4840]: I0930 14:00:47.367849 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c2b4d42-9935-4e56-972f-cdd95d5a0e86" containerName="extract-utilities" Sep 30 14:00:47 crc kubenswrapper[4840]: E0930 14:00:47.367858 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c97b5b98-9a09-4c3b-9bd0-5a5a8671c479" containerName="registry-server" Sep 30 14:00:47 crc kubenswrapper[4840]: I0930 14:00:47.367865 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="c97b5b98-9a09-4c3b-9bd0-5a5a8671c479" containerName="registry-server" Sep 30 14:00:47 crc kubenswrapper[4840]: E0930 14:00:47.367879 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32d505eb-6035-488f-b154-2699e8af36ed" containerName="marketplace-operator" Sep 30 14:00:47 crc kubenswrapper[4840]: I0930 14:00:47.367886 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="32d505eb-6035-488f-b154-2699e8af36ed" containerName="marketplace-operator" Sep 30 14:00:47 crc kubenswrapper[4840]: E0930 14:00:47.367896 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c2b4d42-9935-4e56-972f-cdd95d5a0e86" containerName="extract-content" Sep 30 14:00:47 crc kubenswrapper[4840]: I0930 14:00:47.367903 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c2b4d42-9935-4e56-972f-cdd95d5a0e86" containerName="extract-content" Sep 30 14:00:47 crc kubenswrapper[4840]: E0930 14:00:47.367913 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d587bdc-9a81-4e7c-9df2-d638d43f5e0d" containerName="extract-utilities" Sep 30 14:00:47 crc kubenswrapper[4840]: I0930 14:00:47.367921 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d587bdc-9a81-4e7c-9df2-d638d43f5e0d" containerName="extract-utilities" Sep 30 14:00:47 crc kubenswrapper[4840]: E0930 14:00:47.367930 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d587bdc-9a81-4e7c-9df2-d638d43f5e0d" containerName="registry-server" Sep 30 14:00:47 crc kubenswrapper[4840]: I0930 14:00:47.367937 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d587bdc-9a81-4e7c-9df2-d638d43f5e0d" containerName="registry-server" Sep 30 14:00:47 crc kubenswrapper[4840]: E0930 14:00:47.367950 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f1c8962-d085-430f-9654-9aa798bc399c" containerName="registry-server" Sep 30 14:00:47 crc kubenswrapper[4840]: I0930 14:00:47.367958 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f1c8962-d085-430f-9654-9aa798bc399c" containerName="registry-server" Sep 30 14:00:47 crc kubenswrapper[4840]: E0930 14:00:47.367968 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c2b4d42-9935-4e56-972f-cdd95d5a0e86" containerName="registry-server" Sep 30 14:00:47 crc kubenswrapper[4840]: I0930 14:00:47.367975 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c2b4d42-9935-4e56-972f-cdd95d5a0e86" containerName="registry-server" Sep 30 14:00:47 crc kubenswrapper[4840]: E0930 14:00:47.367987 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d587bdc-9a81-4e7c-9df2-d638d43f5e0d" containerName="extract-content" Sep 30 14:00:47 crc kubenswrapper[4840]: I0930 14:00:47.367994 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d587bdc-9a81-4e7c-9df2-d638d43f5e0d" containerName="extract-content" Sep 30 14:00:47 crc kubenswrapper[4840]: I0930 14:00:47.368107 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="32d505eb-6035-488f-b154-2699e8af36ed" containerName="marketplace-operator" Sep 30 14:00:47 crc kubenswrapper[4840]: I0930 14:00:47.368120 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c2b4d42-9935-4e56-972f-cdd95d5a0e86" containerName="registry-server" Sep 30 14:00:47 crc kubenswrapper[4840]: I0930 14:00:47.368128 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f1c8962-d085-430f-9654-9aa798bc399c" containerName="registry-server" Sep 30 14:00:47 crc kubenswrapper[4840]: I0930 14:00:47.368139 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="c97b5b98-9a09-4c3b-9bd0-5a5a8671c479" containerName="registry-server" Sep 30 14:00:47 crc kubenswrapper[4840]: I0930 14:00:47.368149 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d587bdc-9a81-4e7c-9df2-d638d43f5e0d" containerName="registry-server" Sep 30 14:00:47 crc kubenswrapper[4840]: I0930 14:00:47.369002 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f85rt" Sep 30 14:00:47 crc kubenswrapper[4840]: I0930 14:00:47.370896 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Sep 30 14:00:47 crc kubenswrapper[4840]: I0930 14:00:47.389927 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f85rt"] Sep 30 14:00:47 crc kubenswrapper[4840]: I0930 14:00:47.434020 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crggr\" (UniqueName: \"kubernetes.io/projected/0207a121-1290-431e-85a5-e5bfc5355d50-kube-api-access-crggr\") pod \"redhat-marketplace-f85rt\" (UID: \"0207a121-1290-431e-85a5-e5bfc5355d50\") " pod="openshift-marketplace/redhat-marketplace-f85rt" Sep 30 14:00:47 crc kubenswrapper[4840]: I0930 14:00:47.434098 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0207a121-1290-431e-85a5-e5bfc5355d50-utilities\") pod \"redhat-marketplace-f85rt\" (UID: \"0207a121-1290-431e-85a5-e5bfc5355d50\") " pod="openshift-marketplace/redhat-marketplace-f85rt" Sep 30 14:00:47 crc kubenswrapper[4840]: I0930 14:00:47.434122 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0207a121-1290-431e-85a5-e5bfc5355d50-catalog-content\") pod \"redhat-marketplace-f85rt\" (UID: \"0207a121-1290-431e-85a5-e5bfc5355d50\") " pod="openshift-marketplace/redhat-marketplace-f85rt" Sep 30 14:00:47 crc kubenswrapper[4840]: I0930 14:00:47.535329 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0207a121-1290-431e-85a5-e5bfc5355d50-utilities\") pod \"redhat-marketplace-f85rt\" (UID: \"0207a121-1290-431e-85a5-e5bfc5355d50\") " pod="openshift-marketplace/redhat-marketplace-f85rt" Sep 30 14:00:47 crc kubenswrapper[4840]: I0930 14:00:47.535694 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0207a121-1290-431e-85a5-e5bfc5355d50-catalog-content\") pod \"redhat-marketplace-f85rt\" (UID: \"0207a121-1290-431e-85a5-e5bfc5355d50\") " pod="openshift-marketplace/redhat-marketplace-f85rt" Sep 30 14:00:47 crc kubenswrapper[4840]: I0930 14:00:47.535778 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crggr\" (UniqueName: \"kubernetes.io/projected/0207a121-1290-431e-85a5-e5bfc5355d50-kube-api-access-crggr\") pod \"redhat-marketplace-f85rt\" (UID: \"0207a121-1290-431e-85a5-e5bfc5355d50\") " pod="openshift-marketplace/redhat-marketplace-f85rt" Sep 30 14:00:47 crc kubenswrapper[4840]: I0930 14:00:47.535920 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0207a121-1290-431e-85a5-e5bfc5355d50-utilities\") pod \"redhat-marketplace-f85rt\" (UID: \"0207a121-1290-431e-85a5-e5bfc5355d50\") " pod="openshift-marketplace/redhat-marketplace-f85rt" Sep 30 14:00:47 crc kubenswrapper[4840]: I0930 14:00:47.536181 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0207a121-1290-431e-85a5-e5bfc5355d50-catalog-content\") pod \"redhat-marketplace-f85rt\" (UID: \"0207a121-1290-431e-85a5-e5bfc5355d50\") " pod="openshift-marketplace/redhat-marketplace-f85rt" Sep 30 14:00:47 crc kubenswrapper[4840]: I0930 14:00:47.554353 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crggr\" (UniqueName: \"kubernetes.io/projected/0207a121-1290-431e-85a5-e5bfc5355d50-kube-api-access-crggr\") pod \"redhat-marketplace-f85rt\" (UID: \"0207a121-1290-431e-85a5-e5bfc5355d50\") " pod="openshift-marketplace/redhat-marketplace-f85rt" Sep 30 14:00:47 crc kubenswrapper[4840]: I0930 14:00:47.687368 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f85rt" Sep 30 14:00:47 crc kubenswrapper[4840]: I0930 14:00:47.965940 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4hhpc"] Sep 30 14:00:47 crc kubenswrapper[4840]: I0930 14:00:47.966884 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4hhpc" Sep 30 14:00:47 crc kubenswrapper[4840]: I0930 14:00:47.970162 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Sep 30 14:00:47 crc kubenswrapper[4840]: I0930 14:00:47.972932 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4hhpc"] Sep 30 14:00:48 crc kubenswrapper[4840]: I0930 14:00:48.041947 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2818028-42ff-459b-ab25-0f05227b88fd-catalog-content\") pod \"redhat-operators-4hhpc\" (UID: \"f2818028-42ff-459b-ab25-0f05227b88fd\") " pod="openshift-marketplace/redhat-operators-4hhpc" Sep 30 14:00:48 crc kubenswrapper[4840]: I0930 14:00:48.042259 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l856r\" (UniqueName: \"kubernetes.io/projected/f2818028-42ff-459b-ab25-0f05227b88fd-kube-api-access-l856r\") pod \"redhat-operators-4hhpc\" (UID: \"f2818028-42ff-459b-ab25-0f05227b88fd\") " pod="openshift-marketplace/redhat-operators-4hhpc" Sep 30 14:00:48 crc kubenswrapper[4840]: I0930 14:00:48.042301 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2818028-42ff-459b-ab25-0f05227b88fd-utilities\") pod \"redhat-operators-4hhpc\" (UID: \"f2818028-42ff-459b-ab25-0f05227b88fd\") " pod="openshift-marketplace/redhat-operators-4hhpc" Sep 30 14:00:48 crc kubenswrapper[4840]: I0930 14:00:48.072476 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f85rt"] Sep 30 14:00:48 crc kubenswrapper[4840]: W0930 14:00:48.078586 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0207a121_1290_431e_85a5_e5bfc5355d50.slice/crio-923f8f62c0728674ddb7f0051b55ab1dd97deac1c013d076aa956af08ab971a4 WatchSource:0}: Error finding container 923f8f62c0728674ddb7f0051b55ab1dd97deac1c013d076aa956af08ab971a4: Status 404 returned error can't find the container with id 923f8f62c0728674ddb7f0051b55ab1dd97deac1c013d076aa956af08ab971a4 Sep 30 14:00:48 crc kubenswrapper[4840]: I0930 14:00:48.143905 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2818028-42ff-459b-ab25-0f05227b88fd-catalog-content\") pod \"redhat-operators-4hhpc\" (UID: \"f2818028-42ff-459b-ab25-0f05227b88fd\") " pod="openshift-marketplace/redhat-operators-4hhpc" Sep 30 14:00:48 crc kubenswrapper[4840]: I0930 14:00:48.144048 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l856r\" (UniqueName: \"kubernetes.io/projected/f2818028-42ff-459b-ab25-0f05227b88fd-kube-api-access-l856r\") pod \"redhat-operators-4hhpc\" (UID: \"f2818028-42ff-459b-ab25-0f05227b88fd\") " pod="openshift-marketplace/redhat-operators-4hhpc" Sep 30 14:00:48 crc kubenswrapper[4840]: I0930 14:00:48.144241 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2818028-42ff-459b-ab25-0f05227b88fd-utilities\") pod \"redhat-operators-4hhpc\" (UID: \"f2818028-42ff-459b-ab25-0f05227b88fd\") " pod="openshift-marketplace/redhat-operators-4hhpc" Sep 30 14:00:48 crc kubenswrapper[4840]: I0930 14:00:48.144730 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2818028-42ff-459b-ab25-0f05227b88fd-catalog-content\") pod \"redhat-operators-4hhpc\" (UID: \"f2818028-42ff-459b-ab25-0f05227b88fd\") " pod="openshift-marketplace/redhat-operators-4hhpc" Sep 30 14:00:48 crc kubenswrapper[4840]: I0930 14:00:48.144756 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2818028-42ff-459b-ab25-0f05227b88fd-utilities\") pod \"redhat-operators-4hhpc\" (UID: \"f2818028-42ff-459b-ab25-0f05227b88fd\") " pod="openshift-marketplace/redhat-operators-4hhpc" Sep 30 14:00:48 crc kubenswrapper[4840]: I0930 14:00:48.167059 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l856r\" (UniqueName: \"kubernetes.io/projected/f2818028-42ff-459b-ab25-0f05227b88fd-kube-api-access-l856r\") pod \"redhat-operators-4hhpc\" (UID: \"f2818028-42ff-459b-ab25-0f05227b88fd\") " pod="openshift-marketplace/redhat-operators-4hhpc" Sep 30 14:00:48 crc kubenswrapper[4840]: I0930 14:00:48.281813 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4hhpc" Sep 30 14:00:48 crc kubenswrapper[4840]: I0930 14:00:48.653139 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4hhpc"] Sep 30 14:00:48 crc kubenswrapper[4840]: W0930 14:00:48.659000 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf2818028_42ff_459b_ab25_0f05227b88fd.slice/crio-cdd79db3e5dc511ee741e8d1ac32b6586a588a96981ffe6b25e900eb660221a4 WatchSource:0}: Error finding container cdd79db3e5dc511ee741e8d1ac32b6586a588a96981ffe6b25e900eb660221a4: Status 404 returned error can't find the container with id cdd79db3e5dc511ee741e8d1ac32b6586a588a96981ffe6b25e900eb660221a4 Sep 30 14:00:48 crc kubenswrapper[4840]: I0930 14:00:48.765108 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4hhpc" event={"ID":"f2818028-42ff-459b-ab25-0f05227b88fd","Type":"ContainerStarted","Data":"cdd79db3e5dc511ee741e8d1ac32b6586a588a96981ffe6b25e900eb660221a4"} Sep 30 14:00:48 crc kubenswrapper[4840]: I0930 14:00:48.767702 4840 generic.go:334] "Generic (PLEG): container finished" podID="0207a121-1290-431e-85a5-e5bfc5355d50" containerID="7dfe4a1026ee8c580d558a0baec0cc255ca4fe65e7f92baea9e2cdafebeb6ac6" exitCode=0 Sep 30 14:00:48 crc kubenswrapper[4840]: I0930 14:00:48.768204 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f85rt" event={"ID":"0207a121-1290-431e-85a5-e5bfc5355d50","Type":"ContainerDied","Data":"7dfe4a1026ee8c580d558a0baec0cc255ca4fe65e7f92baea9e2cdafebeb6ac6"} Sep 30 14:00:48 crc kubenswrapper[4840]: I0930 14:00:48.768283 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f85rt" event={"ID":"0207a121-1290-431e-85a5-e5bfc5355d50","Type":"ContainerStarted","Data":"923f8f62c0728674ddb7f0051b55ab1dd97deac1c013d076aa956af08ab971a4"} Sep 30 14:00:49 crc kubenswrapper[4840]: I0930 14:00:49.767163 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zbb8b"] Sep 30 14:00:49 crc kubenswrapper[4840]: I0930 14:00:49.768540 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zbb8b" Sep 30 14:00:49 crc kubenswrapper[4840]: I0930 14:00:49.770455 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Sep 30 14:00:49 crc kubenswrapper[4840]: I0930 14:00:49.774790 4840 generic.go:334] "Generic (PLEG): container finished" podID="f2818028-42ff-459b-ab25-0f05227b88fd" containerID="de0a31658503745458c591d53934c7bb41a9306b3a67314b29f088097f0e6c35" exitCode=0 Sep 30 14:00:49 crc kubenswrapper[4840]: I0930 14:00:49.774872 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4hhpc" event={"ID":"f2818028-42ff-459b-ab25-0f05227b88fd","Type":"ContainerDied","Data":"de0a31658503745458c591d53934c7bb41a9306b3a67314b29f088097f0e6c35"} Sep 30 14:00:49 crc kubenswrapper[4840]: I0930 14:00:49.780157 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f85rt" event={"ID":"0207a121-1290-431e-85a5-e5bfc5355d50","Type":"ContainerStarted","Data":"cead2ac3e953fb7618bde0f8a0f8edb2fa52e3dc8d80cd4ff966cfce631200af"} Sep 30 14:00:49 crc kubenswrapper[4840]: I0930 14:00:49.787104 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zbb8b"] Sep 30 14:00:49 crc kubenswrapper[4840]: I0930 14:00:49.866627 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/233f29d4-5cbe-4cd1-b8ac-e40cf332d381-utilities\") pod \"certified-operators-zbb8b\" (UID: \"233f29d4-5cbe-4cd1-b8ac-e40cf332d381\") " pod="openshift-marketplace/certified-operators-zbb8b" Sep 30 14:00:49 crc kubenswrapper[4840]: I0930 14:00:49.866704 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvssl\" (UniqueName: \"kubernetes.io/projected/233f29d4-5cbe-4cd1-b8ac-e40cf332d381-kube-api-access-vvssl\") pod \"certified-operators-zbb8b\" (UID: \"233f29d4-5cbe-4cd1-b8ac-e40cf332d381\") " pod="openshift-marketplace/certified-operators-zbb8b" Sep 30 14:00:49 crc kubenswrapper[4840]: I0930 14:00:49.866829 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/233f29d4-5cbe-4cd1-b8ac-e40cf332d381-catalog-content\") pod \"certified-operators-zbb8b\" (UID: \"233f29d4-5cbe-4cd1-b8ac-e40cf332d381\") " pod="openshift-marketplace/certified-operators-zbb8b" Sep 30 14:00:49 crc kubenswrapper[4840]: I0930 14:00:49.967717 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/233f29d4-5cbe-4cd1-b8ac-e40cf332d381-catalog-content\") pod \"certified-operators-zbb8b\" (UID: \"233f29d4-5cbe-4cd1-b8ac-e40cf332d381\") " pod="openshift-marketplace/certified-operators-zbb8b" Sep 30 14:00:49 crc kubenswrapper[4840]: I0930 14:00:49.967776 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/233f29d4-5cbe-4cd1-b8ac-e40cf332d381-utilities\") pod \"certified-operators-zbb8b\" (UID: \"233f29d4-5cbe-4cd1-b8ac-e40cf332d381\") " pod="openshift-marketplace/certified-operators-zbb8b" Sep 30 14:00:49 crc kubenswrapper[4840]: I0930 14:00:49.967813 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvssl\" (UniqueName: \"kubernetes.io/projected/233f29d4-5cbe-4cd1-b8ac-e40cf332d381-kube-api-access-vvssl\") pod \"certified-operators-zbb8b\" (UID: \"233f29d4-5cbe-4cd1-b8ac-e40cf332d381\") " pod="openshift-marketplace/certified-operators-zbb8b" Sep 30 14:00:49 crc kubenswrapper[4840]: I0930 14:00:49.968319 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/233f29d4-5cbe-4cd1-b8ac-e40cf332d381-catalog-content\") pod \"certified-operators-zbb8b\" (UID: \"233f29d4-5cbe-4cd1-b8ac-e40cf332d381\") " pod="openshift-marketplace/certified-operators-zbb8b" Sep 30 14:00:49 crc kubenswrapper[4840]: I0930 14:00:49.968338 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/233f29d4-5cbe-4cd1-b8ac-e40cf332d381-utilities\") pod \"certified-operators-zbb8b\" (UID: \"233f29d4-5cbe-4cd1-b8ac-e40cf332d381\") " pod="openshift-marketplace/certified-operators-zbb8b" Sep 30 14:00:49 crc kubenswrapper[4840]: I0930 14:00:49.987515 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvssl\" (UniqueName: \"kubernetes.io/projected/233f29d4-5cbe-4cd1-b8ac-e40cf332d381-kube-api-access-vvssl\") pod \"certified-operators-zbb8b\" (UID: \"233f29d4-5cbe-4cd1-b8ac-e40cf332d381\") " pod="openshift-marketplace/certified-operators-zbb8b" Sep 30 14:00:50 crc kubenswrapper[4840]: I0930 14:00:50.085817 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zbb8b" Sep 30 14:00:50 crc kubenswrapper[4840]: I0930 14:00:50.367746 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-t2tp4"] Sep 30 14:00:50 crc kubenswrapper[4840]: I0930 14:00:50.369071 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t2tp4" Sep 30 14:00:50 crc kubenswrapper[4840]: I0930 14:00:50.377072 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Sep 30 14:00:50 crc kubenswrapper[4840]: I0930 14:00:50.384086 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t2tp4"] Sep 30 14:00:50 crc kubenswrapper[4840]: I0930 14:00:50.471806 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zbb8b"] Sep 30 14:00:50 crc kubenswrapper[4840]: I0930 14:00:50.479351 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5grc\" (UniqueName: \"kubernetes.io/projected/e5a51075-4798-48bf-ac82-67a642d1269a-kube-api-access-t5grc\") pod \"community-operators-t2tp4\" (UID: \"e5a51075-4798-48bf-ac82-67a642d1269a\") " pod="openshift-marketplace/community-operators-t2tp4" Sep 30 14:00:50 crc kubenswrapper[4840]: I0930 14:00:50.479408 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5a51075-4798-48bf-ac82-67a642d1269a-catalog-content\") pod \"community-operators-t2tp4\" (UID: \"e5a51075-4798-48bf-ac82-67a642d1269a\") " pod="openshift-marketplace/community-operators-t2tp4" Sep 30 14:00:50 crc kubenswrapper[4840]: I0930 14:00:50.479429 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5a51075-4798-48bf-ac82-67a642d1269a-utilities\") pod \"community-operators-t2tp4\" (UID: \"e5a51075-4798-48bf-ac82-67a642d1269a\") " pod="openshift-marketplace/community-operators-t2tp4" Sep 30 14:00:50 crc kubenswrapper[4840]: I0930 14:00:50.581367 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5grc\" (UniqueName: \"kubernetes.io/projected/e5a51075-4798-48bf-ac82-67a642d1269a-kube-api-access-t5grc\") pod \"community-operators-t2tp4\" (UID: \"e5a51075-4798-48bf-ac82-67a642d1269a\") " pod="openshift-marketplace/community-operators-t2tp4" Sep 30 14:00:50 crc kubenswrapper[4840]: I0930 14:00:50.581466 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5a51075-4798-48bf-ac82-67a642d1269a-catalog-content\") pod \"community-operators-t2tp4\" (UID: \"e5a51075-4798-48bf-ac82-67a642d1269a\") " pod="openshift-marketplace/community-operators-t2tp4" Sep 30 14:00:50 crc kubenswrapper[4840]: I0930 14:00:50.581493 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5a51075-4798-48bf-ac82-67a642d1269a-utilities\") pod \"community-operators-t2tp4\" (UID: \"e5a51075-4798-48bf-ac82-67a642d1269a\") " pod="openshift-marketplace/community-operators-t2tp4" Sep 30 14:00:50 crc kubenswrapper[4840]: I0930 14:00:50.582137 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5a51075-4798-48bf-ac82-67a642d1269a-utilities\") pod \"community-operators-t2tp4\" (UID: \"e5a51075-4798-48bf-ac82-67a642d1269a\") " pod="openshift-marketplace/community-operators-t2tp4" Sep 30 14:00:50 crc kubenswrapper[4840]: I0930 14:00:50.582262 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5a51075-4798-48bf-ac82-67a642d1269a-catalog-content\") pod \"community-operators-t2tp4\" (UID: \"e5a51075-4798-48bf-ac82-67a642d1269a\") " pod="openshift-marketplace/community-operators-t2tp4" Sep 30 14:00:50 crc kubenswrapper[4840]: I0930 14:00:50.603499 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5grc\" (UniqueName: \"kubernetes.io/projected/e5a51075-4798-48bf-ac82-67a642d1269a-kube-api-access-t5grc\") pod \"community-operators-t2tp4\" (UID: \"e5a51075-4798-48bf-ac82-67a642d1269a\") " pod="openshift-marketplace/community-operators-t2tp4" Sep 30 14:00:50 crc kubenswrapper[4840]: I0930 14:00:50.698565 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t2tp4" Sep 30 14:00:50 crc kubenswrapper[4840]: I0930 14:00:50.786517 4840 generic.go:334] "Generic (PLEG): container finished" podID="233f29d4-5cbe-4cd1-b8ac-e40cf332d381" containerID="01c1e6486070c507c9d3e4342412198c7d7004a22290561a70ef173b83f00fe8" exitCode=0 Sep 30 14:00:50 crc kubenswrapper[4840]: I0930 14:00:50.786588 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zbb8b" event={"ID":"233f29d4-5cbe-4cd1-b8ac-e40cf332d381","Type":"ContainerDied","Data":"01c1e6486070c507c9d3e4342412198c7d7004a22290561a70ef173b83f00fe8"} Sep 30 14:00:50 crc kubenswrapper[4840]: I0930 14:00:50.787039 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zbb8b" event={"ID":"233f29d4-5cbe-4cd1-b8ac-e40cf332d381","Type":"ContainerStarted","Data":"0fb87a16c4ffce2402cf793e428f4b2f75efbcf3a4074c4e397766a1266ce4e9"} Sep 30 14:00:50 crc kubenswrapper[4840]: I0930 14:00:50.788826 4840 generic.go:334] "Generic (PLEG): container finished" podID="0207a121-1290-431e-85a5-e5bfc5355d50" containerID="cead2ac3e953fb7618bde0f8a0f8edb2fa52e3dc8d80cd4ff966cfce631200af" exitCode=0 Sep 30 14:00:50 crc kubenswrapper[4840]: I0930 14:00:50.788871 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f85rt" event={"ID":"0207a121-1290-431e-85a5-e5bfc5355d50","Type":"ContainerDied","Data":"cead2ac3e953fb7618bde0f8a0f8edb2fa52e3dc8d80cd4ff966cfce631200af"} Sep 30 14:00:51 crc kubenswrapper[4840]: I0930 14:00:51.080233 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t2tp4"] Sep 30 14:00:51 crc kubenswrapper[4840]: W0930 14:00:51.087046 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode5a51075_4798_48bf_ac82_67a642d1269a.slice/crio-9759a8a21a3167dffe576b212f26fb1883691c78b2390d18f06d5dec6ea43cd1 WatchSource:0}: Error finding container 9759a8a21a3167dffe576b212f26fb1883691c78b2390d18f06d5dec6ea43cd1: Status 404 returned error can't find the container with id 9759a8a21a3167dffe576b212f26fb1883691c78b2390d18f06d5dec6ea43cd1 Sep 30 14:00:51 crc kubenswrapper[4840]: I0930 14:00:51.798945 4840 generic.go:334] "Generic (PLEG): container finished" podID="f2818028-42ff-459b-ab25-0f05227b88fd" containerID="007d9bb50d7a7033c1a985a9d5baf7f7a4ff2fe9aebd497b095cd58ba32c6592" exitCode=0 Sep 30 14:00:51 crc kubenswrapper[4840]: I0930 14:00:51.799162 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4hhpc" event={"ID":"f2818028-42ff-459b-ab25-0f05227b88fd","Type":"ContainerDied","Data":"007d9bb50d7a7033c1a985a9d5baf7f7a4ff2fe9aebd497b095cd58ba32c6592"} Sep 30 14:00:51 crc kubenswrapper[4840]: I0930 14:00:51.804463 4840 generic.go:334] "Generic (PLEG): container finished" podID="e5a51075-4798-48bf-ac82-67a642d1269a" containerID="c53cb3944ec6b4d13a767827f8e87ea1ff739afdaff921f66def5bfcc47a8b9e" exitCode=0 Sep 30 14:00:51 crc kubenswrapper[4840]: I0930 14:00:51.804502 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t2tp4" event={"ID":"e5a51075-4798-48bf-ac82-67a642d1269a","Type":"ContainerDied","Data":"c53cb3944ec6b4d13a767827f8e87ea1ff739afdaff921f66def5bfcc47a8b9e"} Sep 30 14:00:51 crc kubenswrapper[4840]: I0930 14:00:51.804664 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t2tp4" event={"ID":"e5a51075-4798-48bf-ac82-67a642d1269a","Type":"ContainerStarted","Data":"9759a8a21a3167dffe576b212f26fb1883691c78b2390d18f06d5dec6ea43cd1"} Sep 30 14:00:52 crc kubenswrapper[4840]: I0930 14:00:52.811064 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zbb8b" event={"ID":"233f29d4-5cbe-4cd1-b8ac-e40cf332d381","Type":"ContainerStarted","Data":"75ae6429f8e97db264eb05ac961b5f62c172dfdaa0d40e643324ab0b9453e3fe"} Sep 30 14:00:52 crc kubenswrapper[4840]: I0930 14:00:52.812705 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t2tp4" event={"ID":"e5a51075-4798-48bf-ac82-67a642d1269a","Type":"ContainerStarted","Data":"8f002688f13a0db9631486a4b1602411ec8f5de05c142f1a40741b881204a71c"} Sep 30 14:00:52 crc kubenswrapper[4840]: I0930 14:00:52.814605 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f85rt" event={"ID":"0207a121-1290-431e-85a5-e5bfc5355d50","Type":"ContainerStarted","Data":"6452b33ff327f7100b2d34d8f960f49259858e9e62816cb748dbdfc4b42c028e"} Sep 30 14:00:52 crc kubenswrapper[4840]: I0930 14:00:52.870607 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-f85rt" podStartSLOduration=2.943537835 podStartE2EDuration="5.870573826s" podCreationTimestamp="2025-09-30 14:00:47 +0000 UTC" firstStartedPulling="2025-09-30 14:00:48.7697589 +0000 UTC m=+277.398845313" lastFinishedPulling="2025-09-30 14:00:51.696794881 +0000 UTC m=+280.325881304" observedRunningTime="2025-09-30 14:00:52.868312996 +0000 UTC m=+281.497399429" watchObservedRunningTime="2025-09-30 14:00:52.870573826 +0000 UTC m=+281.499660249" Sep 30 14:00:53 crc kubenswrapper[4840]: I0930 14:00:53.821563 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4hhpc" event={"ID":"f2818028-42ff-459b-ab25-0f05227b88fd","Type":"ContainerStarted","Data":"ce79b7653d94d39c8257d5408f797f1a19beae2326136d5522df025cc846b0c4"} Sep 30 14:00:53 crc kubenswrapper[4840]: I0930 14:00:53.824408 4840 generic.go:334] "Generic (PLEG): container finished" podID="e5a51075-4798-48bf-ac82-67a642d1269a" containerID="8f002688f13a0db9631486a4b1602411ec8f5de05c142f1a40741b881204a71c" exitCode=0 Sep 30 14:00:53 crc kubenswrapper[4840]: I0930 14:00:53.824470 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t2tp4" event={"ID":"e5a51075-4798-48bf-ac82-67a642d1269a","Type":"ContainerDied","Data":"8f002688f13a0db9631486a4b1602411ec8f5de05c142f1a40741b881204a71c"} Sep 30 14:00:53 crc kubenswrapper[4840]: I0930 14:00:53.826193 4840 generic.go:334] "Generic (PLEG): container finished" podID="233f29d4-5cbe-4cd1-b8ac-e40cf332d381" containerID="75ae6429f8e97db264eb05ac961b5f62c172dfdaa0d40e643324ab0b9453e3fe" exitCode=0 Sep 30 14:00:53 crc kubenswrapper[4840]: I0930 14:00:53.827006 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zbb8b" event={"ID":"233f29d4-5cbe-4cd1-b8ac-e40cf332d381","Type":"ContainerDied","Data":"75ae6429f8e97db264eb05ac961b5f62c172dfdaa0d40e643324ab0b9453e3fe"} Sep 30 14:00:53 crc kubenswrapper[4840]: I0930 14:00:53.840158 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4hhpc" podStartSLOduration=3.931149477 podStartE2EDuration="6.840141187s" podCreationTimestamp="2025-09-30 14:00:47 +0000 UTC" firstStartedPulling="2025-09-30 14:00:49.775954667 +0000 UTC m=+278.405041090" lastFinishedPulling="2025-09-30 14:00:52.684946377 +0000 UTC m=+281.314032800" observedRunningTime="2025-09-30 14:00:53.838032601 +0000 UTC m=+282.467119024" watchObservedRunningTime="2025-09-30 14:00:53.840141187 +0000 UTC m=+282.469227610" Sep 30 14:00:54 crc kubenswrapper[4840]: I0930 14:00:54.837406 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zbb8b" event={"ID":"233f29d4-5cbe-4cd1-b8ac-e40cf332d381","Type":"ContainerStarted","Data":"5bac98b23fbc99288dee0d78371ed7ebc6fcfab8d08302da88ba6667db7c2ba6"} Sep 30 14:00:54 crc kubenswrapper[4840]: I0930 14:00:54.841516 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t2tp4" event={"ID":"e5a51075-4798-48bf-ac82-67a642d1269a","Type":"ContainerStarted","Data":"3210f4d8b25c37367beb5eb6807e2586568932e60d098eaa2175ae7fb13b0a98"} Sep 30 14:00:54 crc kubenswrapper[4840]: I0930 14:00:54.855460 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zbb8b" podStartSLOduration=2.371163129 podStartE2EDuration="5.855442507s" podCreationTimestamp="2025-09-30 14:00:49 +0000 UTC" firstStartedPulling="2025-09-30 14:00:50.788006111 +0000 UTC m=+279.417092534" lastFinishedPulling="2025-09-30 14:00:54.272285489 +0000 UTC m=+282.901371912" observedRunningTime="2025-09-30 14:00:54.854222545 +0000 UTC m=+283.483308968" watchObservedRunningTime="2025-09-30 14:00:54.855442507 +0000 UTC m=+283.484528930" Sep 30 14:00:54 crc kubenswrapper[4840]: I0930 14:00:54.869216 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-t2tp4" podStartSLOduration=2.168319053 podStartE2EDuration="4.869200004s" podCreationTimestamp="2025-09-30 14:00:50 +0000 UTC" firstStartedPulling="2025-09-30 14:00:51.805965312 +0000 UTC m=+280.435051745" lastFinishedPulling="2025-09-30 14:00:54.506846273 +0000 UTC m=+283.135932696" observedRunningTime="2025-09-30 14:00:54.868970148 +0000 UTC m=+283.498056571" watchObservedRunningTime="2025-09-30 14:00:54.869200004 +0000 UTC m=+283.498286427" Sep 30 14:00:57 crc kubenswrapper[4840]: I0930 14:00:57.687908 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-f85rt" Sep 30 14:00:57 crc kubenswrapper[4840]: I0930 14:00:57.688252 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-f85rt" Sep 30 14:00:57 crc kubenswrapper[4840]: I0930 14:00:57.726264 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-f85rt" Sep 30 14:00:57 crc kubenswrapper[4840]: I0930 14:00:57.906834 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-f85rt" Sep 30 14:00:58 crc kubenswrapper[4840]: I0930 14:00:58.282313 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4hhpc" Sep 30 14:00:58 crc kubenswrapper[4840]: I0930 14:00:58.282609 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4hhpc" Sep 30 14:00:58 crc kubenswrapper[4840]: I0930 14:00:58.326464 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4hhpc" Sep 30 14:00:58 crc kubenswrapper[4840]: I0930 14:00:58.925082 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4hhpc" Sep 30 14:01:00 crc kubenswrapper[4840]: I0930 14:01:00.086968 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zbb8b" Sep 30 14:01:00 crc kubenswrapper[4840]: I0930 14:01:00.087015 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zbb8b" Sep 30 14:01:00 crc kubenswrapper[4840]: I0930 14:01:00.128668 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zbb8b" Sep 30 14:01:00 crc kubenswrapper[4840]: I0930 14:01:00.700298 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-t2tp4" Sep 30 14:01:00 crc kubenswrapper[4840]: I0930 14:01:00.700350 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-t2tp4" Sep 30 14:01:00 crc kubenswrapper[4840]: I0930 14:01:00.736419 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-t2tp4" Sep 30 14:01:00 crc kubenswrapper[4840]: I0930 14:01:00.917202 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zbb8b" Sep 30 14:01:00 crc kubenswrapper[4840]: I0930 14:01:00.919667 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-t2tp4" Sep 30 14:02:21 crc kubenswrapper[4840]: I0930 14:02:21.872279 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:02:21 crc kubenswrapper[4840]: I0930 14:02:21.872814 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:02:51 crc kubenswrapper[4840]: I0930 14:02:51.871461 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:02:51 crc kubenswrapper[4840]: I0930 14:02:51.871889 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:03:13 crc kubenswrapper[4840]: I0930 14:03:13.731518 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-x62w9"] Sep 30 14:03:13 crc kubenswrapper[4840]: I0930 14:03:13.732750 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-x62w9" Sep 30 14:03:13 crc kubenswrapper[4840]: I0930 14:03:13.751357 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-x62w9"] Sep 30 14:03:13 crc kubenswrapper[4840]: I0930 14:03:13.803504 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5qc6\" (UniqueName: \"kubernetes.io/projected/357943a7-525f-4f8b-a2d0-31da5a1594e8-kube-api-access-w5qc6\") pod \"image-registry-66df7c8f76-x62w9\" (UID: \"357943a7-525f-4f8b-a2d0-31da5a1594e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-x62w9" Sep 30 14:03:13 crc kubenswrapper[4840]: I0930 14:03:13.803598 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/357943a7-525f-4f8b-a2d0-31da5a1594e8-trusted-ca\") pod \"image-registry-66df7c8f76-x62w9\" (UID: \"357943a7-525f-4f8b-a2d0-31da5a1594e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-x62w9" Sep 30 14:03:13 crc kubenswrapper[4840]: I0930 14:03:13.803626 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/357943a7-525f-4f8b-a2d0-31da5a1594e8-registry-certificates\") pod \"image-registry-66df7c8f76-x62w9\" (UID: \"357943a7-525f-4f8b-a2d0-31da5a1594e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-x62w9" Sep 30 14:03:13 crc kubenswrapper[4840]: I0930 14:03:13.803694 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/357943a7-525f-4f8b-a2d0-31da5a1594e8-installation-pull-secrets\") pod \"image-registry-66df7c8f76-x62w9\" (UID: \"357943a7-525f-4f8b-a2d0-31da5a1594e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-x62w9" Sep 30 14:03:13 crc kubenswrapper[4840]: I0930 14:03:13.803727 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/357943a7-525f-4f8b-a2d0-31da5a1594e8-ca-trust-extracted\") pod \"image-registry-66df7c8f76-x62w9\" (UID: \"357943a7-525f-4f8b-a2d0-31da5a1594e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-x62w9" Sep 30 14:03:13 crc kubenswrapper[4840]: I0930 14:03:13.803756 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/357943a7-525f-4f8b-a2d0-31da5a1594e8-bound-sa-token\") pod \"image-registry-66df7c8f76-x62w9\" (UID: \"357943a7-525f-4f8b-a2d0-31da5a1594e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-x62w9" Sep 30 14:03:13 crc kubenswrapper[4840]: I0930 14:03:13.803798 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-x62w9\" (UID: \"357943a7-525f-4f8b-a2d0-31da5a1594e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-x62w9" Sep 30 14:03:13 crc kubenswrapper[4840]: I0930 14:03:13.803821 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/357943a7-525f-4f8b-a2d0-31da5a1594e8-registry-tls\") pod \"image-registry-66df7c8f76-x62w9\" (UID: \"357943a7-525f-4f8b-a2d0-31da5a1594e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-x62w9" Sep 30 14:03:13 crc kubenswrapper[4840]: I0930 14:03:13.860105 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-x62w9\" (UID: \"357943a7-525f-4f8b-a2d0-31da5a1594e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-x62w9" Sep 30 14:03:13 crc kubenswrapper[4840]: I0930 14:03:13.904972 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/357943a7-525f-4f8b-a2d0-31da5a1594e8-installation-pull-secrets\") pod \"image-registry-66df7c8f76-x62w9\" (UID: \"357943a7-525f-4f8b-a2d0-31da5a1594e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-x62w9" Sep 30 14:03:13 crc kubenswrapper[4840]: I0930 14:03:13.905021 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/357943a7-525f-4f8b-a2d0-31da5a1594e8-ca-trust-extracted\") pod \"image-registry-66df7c8f76-x62w9\" (UID: \"357943a7-525f-4f8b-a2d0-31da5a1594e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-x62w9" Sep 30 14:03:13 crc kubenswrapper[4840]: I0930 14:03:13.905043 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/357943a7-525f-4f8b-a2d0-31da5a1594e8-bound-sa-token\") pod \"image-registry-66df7c8f76-x62w9\" (UID: \"357943a7-525f-4f8b-a2d0-31da5a1594e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-x62w9" Sep 30 14:03:13 crc kubenswrapper[4840]: I0930 14:03:13.905072 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/357943a7-525f-4f8b-a2d0-31da5a1594e8-registry-tls\") pod \"image-registry-66df7c8f76-x62w9\" (UID: \"357943a7-525f-4f8b-a2d0-31da5a1594e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-x62w9" Sep 30 14:03:13 crc kubenswrapper[4840]: I0930 14:03:13.905124 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5qc6\" (UniqueName: \"kubernetes.io/projected/357943a7-525f-4f8b-a2d0-31da5a1594e8-kube-api-access-w5qc6\") pod \"image-registry-66df7c8f76-x62w9\" (UID: \"357943a7-525f-4f8b-a2d0-31da5a1594e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-x62w9" Sep 30 14:03:13 crc kubenswrapper[4840]: I0930 14:03:13.905151 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/357943a7-525f-4f8b-a2d0-31da5a1594e8-trusted-ca\") pod \"image-registry-66df7c8f76-x62w9\" (UID: \"357943a7-525f-4f8b-a2d0-31da5a1594e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-x62w9" Sep 30 14:03:13 crc kubenswrapper[4840]: I0930 14:03:13.905168 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/357943a7-525f-4f8b-a2d0-31da5a1594e8-registry-certificates\") pod \"image-registry-66df7c8f76-x62w9\" (UID: \"357943a7-525f-4f8b-a2d0-31da5a1594e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-x62w9" Sep 30 14:03:13 crc kubenswrapper[4840]: I0930 14:03:13.905728 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/357943a7-525f-4f8b-a2d0-31da5a1594e8-ca-trust-extracted\") pod \"image-registry-66df7c8f76-x62w9\" (UID: \"357943a7-525f-4f8b-a2d0-31da5a1594e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-x62w9" Sep 30 14:03:13 crc kubenswrapper[4840]: I0930 14:03:13.906342 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/357943a7-525f-4f8b-a2d0-31da5a1594e8-trusted-ca\") pod \"image-registry-66df7c8f76-x62w9\" (UID: \"357943a7-525f-4f8b-a2d0-31da5a1594e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-x62w9" Sep 30 14:03:13 crc kubenswrapper[4840]: I0930 14:03:13.906424 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/357943a7-525f-4f8b-a2d0-31da5a1594e8-registry-certificates\") pod \"image-registry-66df7c8f76-x62w9\" (UID: \"357943a7-525f-4f8b-a2d0-31da5a1594e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-x62w9" Sep 30 14:03:13 crc kubenswrapper[4840]: I0930 14:03:13.910964 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/357943a7-525f-4f8b-a2d0-31da5a1594e8-installation-pull-secrets\") pod \"image-registry-66df7c8f76-x62w9\" (UID: \"357943a7-525f-4f8b-a2d0-31da5a1594e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-x62w9" Sep 30 14:03:13 crc kubenswrapper[4840]: I0930 14:03:13.911186 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/357943a7-525f-4f8b-a2d0-31da5a1594e8-registry-tls\") pod \"image-registry-66df7c8f76-x62w9\" (UID: \"357943a7-525f-4f8b-a2d0-31da5a1594e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-x62w9" Sep 30 14:03:13 crc kubenswrapper[4840]: I0930 14:03:13.920504 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/357943a7-525f-4f8b-a2d0-31da5a1594e8-bound-sa-token\") pod \"image-registry-66df7c8f76-x62w9\" (UID: \"357943a7-525f-4f8b-a2d0-31da5a1594e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-x62w9" Sep 30 14:03:13 crc kubenswrapper[4840]: I0930 14:03:13.920902 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5qc6\" (UniqueName: \"kubernetes.io/projected/357943a7-525f-4f8b-a2d0-31da5a1594e8-kube-api-access-w5qc6\") pod \"image-registry-66df7c8f76-x62w9\" (UID: \"357943a7-525f-4f8b-a2d0-31da5a1594e8\") " pod="openshift-image-registry/image-registry-66df7c8f76-x62w9" Sep 30 14:03:14 crc kubenswrapper[4840]: I0930 14:03:14.047001 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-x62w9" Sep 30 14:03:14 crc kubenswrapper[4840]: I0930 14:03:14.245539 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-x62w9"] Sep 30 14:03:14 crc kubenswrapper[4840]: I0930 14:03:14.603183 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-x62w9" event={"ID":"357943a7-525f-4f8b-a2d0-31da5a1594e8","Type":"ContainerStarted","Data":"260edb8bd775ae0c2179acab7ae9b7741fedc1343756363d36d7bce2aca09ebc"} Sep 30 14:03:14 crc kubenswrapper[4840]: I0930 14:03:14.603226 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-x62w9" event={"ID":"357943a7-525f-4f8b-a2d0-31da5a1594e8","Type":"ContainerStarted","Data":"c7e5680e788f7245dc4d5fe20927754d3d91a8c5471a739395ed433a063e1c63"} Sep 30 14:03:14 crc kubenswrapper[4840]: I0930 14:03:14.603295 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-x62w9" Sep 30 14:03:14 crc kubenswrapper[4840]: I0930 14:03:14.626039 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-x62w9" podStartSLOduration=1.626017748 podStartE2EDuration="1.626017748s" podCreationTimestamp="2025-09-30 14:03:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:03:14.622123054 +0000 UTC m=+423.251209487" watchObservedRunningTime="2025-09-30 14:03:14.626017748 +0000 UTC m=+423.255104191" Sep 30 14:03:21 crc kubenswrapper[4840]: I0930 14:03:21.872356 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:03:21 crc kubenswrapper[4840]: I0930 14:03:21.872962 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:03:21 crc kubenswrapper[4840]: I0930 14:03:21.873011 4840 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-747gk" Sep 30 14:03:21 crc kubenswrapper[4840]: I0930 14:03:21.873612 4840 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e98140cf872382a187fd61e08bfb9d9a34afdc7cbc64b4d05d64d64c86bc9322"} pod="openshift-machine-config-operator/machine-config-daemon-747gk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 14:03:21 crc kubenswrapper[4840]: I0930 14:03:21.873677 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" containerID="cri-o://e98140cf872382a187fd61e08bfb9d9a34afdc7cbc64b4d05d64d64c86bc9322" gracePeriod=600 Sep 30 14:03:22 crc kubenswrapper[4840]: I0930 14:03:22.648606 4840 generic.go:334] "Generic (PLEG): container finished" podID="10e8b890-7f20-4a36-8e03-898620cf599a" containerID="e98140cf872382a187fd61e08bfb9d9a34afdc7cbc64b4d05d64d64c86bc9322" exitCode=0 Sep 30 14:03:22 crc kubenswrapper[4840]: I0930 14:03:22.648876 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" event={"ID":"10e8b890-7f20-4a36-8e03-898620cf599a","Type":"ContainerDied","Data":"e98140cf872382a187fd61e08bfb9d9a34afdc7cbc64b4d05d64d64c86bc9322"} Sep 30 14:03:22 crc kubenswrapper[4840]: I0930 14:03:22.649355 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" event={"ID":"10e8b890-7f20-4a36-8e03-898620cf599a","Type":"ContainerStarted","Data":"d12912f690026de70f88356c602b1439b6380a977591bd20b79ac057d9633343"} Sep 30 14:03:22 crc kubenswrapper[4840]: I0930 14:03:22.649393 4840 scope.go:117] "RemoveContainer" containerID="ed16e49c77a448f12cb54d0bc92870e43c4a086fa7add21959f279b2cb557fcd" Sep 30 14:03:34 crc kubenswrapper[4840]: I0930 14:03:34.052452 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-x62w9" Sep 30 14:03:34 crc kubenswrapper[4840]: I0930 14:03:34.101539 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-lzwls"] Sep 30 14:03:59 crc kubenswrapper[4840]: I0930 14:03:59.139361 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" podUID="2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067" containerName="registry" containerID="cri-o://ed46ee312f0c4b2993e82af345239da7c8960d19c03f6c04eb1b8ff14543fb43" gracePeriod=30 Sep 30 14:03:59 crc kubenswrapper[4840]: I0930 14:03:59.478516 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 14:03:59 crc kubenswrapper[4840]: I0930 14:03:59.604001 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-registry-certificates\") pod \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " Sep 30 14:03:59 crc kubenswrapper[4840]: I0930 14:03:59.604364 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " Sep 30 14:03:59 crc kubenswrapper[4840]: I0930 14:03:59.604453 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-registry-tls\") pod \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " Sep 30 14:03:59 crc kubenswrapper[4840]: I0930 14:03:59.604508 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-trusted-ca\") pod \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " Sep 30 14:03:59 crc kubenswrapper[4840]: I0930 14:03:59.604570 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-installation-pull-secrets\") pod \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " Sep 30 14:03:59 crc kubenswrapper[4840]: I0930 14:03:59.604612 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgclp\" (UniqueName: \"kubernetes.io/projected/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-kube-api-access-fgclp\") pod \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " Sep 30 14:03:59 crc kubenswrapper[4840]: I0930 14:03:59.604653 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-bound-sa-token\") pod \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " Sep 30 14:03:59 crc kubenswrapper[4840]: I0930 14:03:59.604741 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-ca-trust-extracted\") pod \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\" (UID: \"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067\") " Sep 30 14:03:59 crc kubenswrapper[4840]: I0930 14:03:59.605418 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:03:59 crc kubenswrapper[4840]: I0930 14:03:59.605606 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:03:59 crc kubenswrapper[4840]: I0930 14:03:59.606069 4840 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 30 14:03:59 crc kubenswrapper[4840]: I0930 14:03:59.606104 4840 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-registry-certificates\") on node \"crc\" DevicePath \"\"" Sep 30 14:03:59 crc kubenswrapper[4840]: I0930 14:03:59.610714 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-kube-api-access-fgclp" (OuterVolumeSpecName: "kube-api-access-fgclp") pod "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067"). InnerVolumeSpecName "kube-api-access-fgclp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:03:59 crc kubenswrapper[4840]: I0930 14:03:59.610940 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:03:59 crc kubenswrapper[4840]: I0930 14:03:59.611136 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:03:59 crc kubenswrapper[4840]: I0930 14:03:59.611297 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:03:59 crc kubenswrapper[4840]: I0930 14:03:59.613850 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Sep 30 14:03:59 crc kubenswrapper[4840]: I0930 14:03:59.625743 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067" (UID: "2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:03:59 crc kubenswrapper[4840]: I0930 14:03:59.707328 4840 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Sep 30 14:03:59 crc kubenswrapper[4840]: I0930 14:03:59.707370 4840 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-registry-tls\") on node \"crc\" DevicePath \"\"" Sep 30 14:03:59 crc kubenswrapper[4840]: I0930 14:03:59.707379 4840 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Sep 30 14:03:59 crc kubenswrapper[4840]: I0930 14:03:59.707390 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgclp\" (UniqueName: \"kubernetes.io/projected/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-kube-api-access-fgclp\") on node \"crc\" DevicePath \"\"" Sep 30 14:03:59 crc kubenswrapper[4840]: I0930 14:03:59.707400 4840 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067-bound-sa-token\") on node \"crc\" DevicePath \"\"" Sep 30 14:03:59 crc kubenswrapper[4840]: I0930 14:03:59.863041 4840 generic.go:334] "Generic (PLEG): container finished" podID="2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067" containerID="ed46ee312f0c4b2993e82af345239da7c8960d19c03f6c04eb1b8ff14543fb43" exitCode=0 Sep 30 14:03:59 crc kubenswrapper[4840]: I0930 14:03:59.863155 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" event={"ID":"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067","Type":"ContainerDied","Data":"ed46ee312f0c4b2993e82af345239da7c8960d19c03f6c04eb1b8ff14543fb43"} Sep 30 14:03:59 crc kubenswrapper[4840]: I0930 14:03:59.863210 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" event={"ID":"2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067","Type":"ContainerDied","Data":"551e02e90f60f1e91c25acaf9d3e8f7b4aa8415f17aabd5be2e1641caabb54ca"} Sep 30 14:03:59 crc kubenswrapper[4840]: I0930 14:03:59.863321 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-lzwls" Sep 30 14:03:59 crc kubenswrapper[4840]: I0930 14:03:59.863347 4840 scope.go:117] "RemoveContainer" containerID="ed46ee312f0c4b2993e82af345239da7c8960d19c03f6c04eb1b8ff14543fb43" Sep 30 14:03:59 crc kubenswrapper[4840]: I0930 14:03:59.882292 4840 scope.go:117] "RemoveContainer" containerID="ed46ee312f0c4b2993e82af345239da7c8960d19c03f6c04eb1b8ff14543fb43" Sep 30 14:03:59 crc kubenswrapper[4840]: E0930 14:03:59.882736 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed46ee312f0c4b2993e82af345239da7c8960d19c03f6c04eb1b8ff14543fb43\": container with ID starting with ed46ee312f0c4b2993e82af345239da7c8960d19c03f6c04eb1b8ff14543fb43 not found: ID does not exist" containerID="ed46ee312f0c4b2993e82af345239da7c8960d19c03f6c04eb1b8ff14543fb43" Sep 30 14:03:59 crc kubenswrapper[4840]: I0930 14:03:59.882824 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed46ee312f0c4b2993e82af345239da7c8960d19c03f6c04eb1b8ff14543fb43"} err="failed to get container status \"ed46ee312f0c4b2993e82af345239da7c8960d19c03f6c04eb1b8ff14543fb43\": rpc error: code = NotFound desc = could not find container \"ed46ee312f0c4b2993e82af345239da7c8960d19c03f6c04eb1b8ff14543fb43\": container with ID starting with ed46ee312f0c4b2993e82af345239da7c8960d19c03f6c04eb1b8ff14543fb43 not found: ID does not exist" Sep 30 14:03:59 crc kubenswrapper[4840]: I0930 14:03:59.908467 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-lzwls"] Sep 30 14:03:59 crc kubenswrapper[4840]: I0930 14:03:59.915527 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-lzwls"] Sep 30 14:04:00 crc kubenswrapper[4840]: I0930 14:04:00.124878 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067" path="/var/lib/kubelet/pods/2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067/volumes" Sep 30 14:05:51 crc kubenswrapper[4840]: I0930 14:05:51.873160 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:05:51 crc kubenswrapper[4840]: I0930 14:05:51.873884 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:06:09 crc kubenswrapper[4840]: I0930 14:06:09.087957 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-f4kpb"] Sep 30 14:06:09 crc kubenswrapper[4840]: E0930 14:06:09.088671 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067" containerName="registry" Sep 30 14:06:09 crc kubenswrapper[4840]: I0930 14:06:09.088684 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067" containerName="registry" Sep 30 14:06:09 crc kubenswrapper[4840]: I0930 14:06:09.088777 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="2eeb9ebb-ba18-4fa7-ac6f-e5cc38cc2067" containerName="registry" Sep 30 14:06:09 crc kubenswrapper[4840]: I0930 14:06:09.089114 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-f4kpb" Sep 30 14:06:09 crc kubenswrapper[4840]: I0930 14:06:09.098891 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-hjfs2"] Sep 30 14:06:09 crc kubenswrapper[4840]: I0930 14:06:09.099946 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-hjfs2" Sep 30 14:06:09 crc kubenswrapper[4840]: I0930 14:06:09.100713 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Sep 30 14:06:09 crc kubenswrapper[4840]: I0930 14:06:09.100938 4840 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-h5hsg" Sep 30 14:06:09 crc kubenswrapper[4840]: I0930 14:06:09.101004 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Sep 30 14:06:09 crc kubenswrapper[4840]: I0930 14:06:09.103611 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-f4kpb"] Sep 30 14:06:09 crc kubenswrapper[4840]: I0930 14:06:09.105888 4840 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-smgdv" Sep 30 14:06:09 crc kubenswrapper[4840]: I0930 14:06:09.121608 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-hjfs2"] Sep 30 14:06:09 crc kubenswrapper[4840]: I0930 14:06:09.125100 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-sszwr"] Sep 30 14:06:09 crc kubenswrapper[4840]: I0930 14:06:09.125905 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-sszwr" Sep 30 14:06:09 crc kubenswrapper[4840]: I0930 14:06:09.138161 4840 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-6dj7g" Sep 30 14:06:09 crc kubenswrapper[4840]: I0930 14:06:09.144008 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-sszwr"] Sep 30 14:06:09 crc kubenswrapper[4840]: I0930 14:06:09.209308 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjqlg\" (UniqueName: \"kubernetes.io/projected/8df1331d-54ff-47cd-80c9-c18a5ab2df53-kube-api-access-vjqlg\") pod \"cert-manager-cainjector-7f985d654d-f4kpb\" (UID: \"8df1331d-54ff-47cd-80c9-c18a5ab2df53\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-f4kpb" Sep 30 14:06:09 crc kubenswrapper[4840]: I0930 14:06:09.209429 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmvzs\" (UniqueName: \"kubernetes.io/projected/11b92f8e-b36f-47e5-8d1a-d3c83ccbd870-kube-api-access-dmvzs\") pod \"cert-manager-5b446d88c5-hjfs2\" (UID: \"11b92f8e-b36f-47e5-8d1a-d3c83ccbd870\") " pod="cert-manager/cert-manager-5b446d88c5-hjfs2" Sep 30 14:06:09 crc kubenswrapper[4840]: I0930 14:06:09.311314 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gp8wp\" (UniqueName: \"kubernetes.io/projected/18d21186-8397-400e-a9df-5e4e41e6739e-kube-api-access-gp8wp\") pod \"cert-manager-webhook-5655c58dd6-sszwr\" (UID: \"18d21186-8397-400e-a9df-5e4e41e6739e\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-sszwr" Sep 30 14:06:09 crc kubenswrapper[4840]: I0930 14:06:09.311813 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjqlg\" (UniqueName: \"kubernetes.io/projected/8df1331d-54ff-47cd-80c9-c18a5ab2df53-kube-api-access-vjqlg\") pod \"cert-manager-cainjector-7f985d654d-f4kpb\" (UID: \"8df1331d-54ff-47cd-80c9-c18a5ab2df53\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-f4kpb" Sep 30 14:06:09 crc kubenswrapper[4840]: I0930 14:06:09.311945 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmvzs\" (UniqueName: \"kubernetes.io/projected/11b92f8e-b36f-47e5-8d1a-d3c83ccbd870-kube-api-access-dmvzs\") pod \"cert-manager-5b446d88c5-hjfs2\" (UID: \"11b92f8e-b36f-47e5-8d1a-d3c83ccbd870\") " pod="cert-manager/cert-manager-5b446d88c5-hjfs2" Sep 30 14:06:09 crc kubenswrapper[4840]: I0930 14:06:09.335507 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjqlg\" (UniqueName: \"kubernetes.io/projected/8df1331d-54ff-47cd-80c9-c18a5ab2df53-kube-api-access-vjqlg\") pod \"cert-manager-cainjector-7f985d654d-f4kpb\" (UID: \"8df1331d-54ff-47cd-80c9-c18a5ab2df53\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-f4kpb" Sep 30 14:06:09 crc kubenswrapper[4840]: I0930 14:06:09.335648 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmvzs\" (UniqueName: \"kubernetes.io/projected/11b92f8e-b36f-47e5-8d1a-d3c83ccbd870-kube-api-access-dmvzs\") pod \"cert-manager-5b446d88c5-hjfs2\" (UID: \"11b92f8e-b36f-47e5-8d1a-d3c83ccbd870\") " pod="cert-manager/cert-manager-5b446d88c5-hjfs2" Sep 30 14:06:09 crc kubenswrapper[4840]: I0930 14:06:09.405041 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-f4kpb" Sep 30 14:06:09 crc kubenswrapper[4840]: I0930 14:06:09.413448 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gp8wp\" (UniqueName: \"kubernetes.io/projected/18d21186-8397-400e-a9df-5e4e41e6739e-kube-api-access-gp8wp\") pod \"cert-manager-webhook-5655c58dd6-sszwr\" (UID: \"18d21186-8397-400e-a9df-5e4e41e6739e\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-sszwr" Sep 30 14:06:09 crc kubenswrapper[4840]: I0930 14:06:09.415224 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-hjfs2" Sep 30 14:06:09 crc kubenswrapper[4840]: I0930 14:06:09.433430 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gp8wp\" (UniqueName: \"kubernetes.io/projected/18d21186-8397-400e-a9df-5e4e41e6739e-kube-api-access-gp8wp\") pod \"cert-manager-webhook-5655c58dd6-sszwr\" (UID: \"18d21186-8397-400e-a9df-5e4e41e6739e\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-sszwr" Sep 30 14:06:09 crc kubenswrapper[4840]: I0930 14:06:09.450634 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-sszwr" Sep 30 14:06:09 crc kubenswrapper[4840]: I0930 14:06:09.661014 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-f4kpb"] Sep 30 14:06:09 crc kubenswrapper[4840]: I0930 14:06:09.681714 4840 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 14:06:09 crc kubenswrapper[4840]: I0930 14:06:09.707490 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-sszwr"] Sep 30 14:06:09 crc kubenswrapper[4840]: W0930 14:06:09.711873 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod18d21186_8397_400e_a9df_5e4e41e6739e.slice/crio-ae7bf5fcd0d76d9b6126c1bb5e75b8e757bc77bd2eb5bc6e25bd23b96e2f72b3 WatchSource:0}: Error finding container ae7bf5fcd0d76d9b6126c1bb5e75b8e757bc77bd2eb5bc6e25bd23b96e2f72b3: Status 404 returned error can't find the container with id ae7bf5fcd0d76d9b6126c1bb5e75b8e757bc77bd2eb5bc6e25bd23b96e2f72b3 Sep 30 14:06:09 crc kubenswrapper[4840]: I0930 14:06:09.724037 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-hjfs2"] Sep 30 14:06:10 crc kubenswrapper[4840]: I0930 14:06:10.547390 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-f4kpb" event={"ID":"8df1331d-54ff-47cd-80c9-c18a5ab2df53","Type":"ContainerStarted","Data":"743a8510678510d78d95818a8455990b360a5bc233fdfd6b54635fb98f2dbb9f"} Sep 30 14:06:10 crc kubenswrapper[4840]: I0930 14:06:10.549518 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-hjfs2" event={"ID":"11b92f8e-b36f-47e5-8d1a-d3c83ccbd870","Type":"ContainerStarted","Data":"99eeb83ed514480c233456b10ce1bcb34a83c8b5ce3eb097585779fd4e497e66"} Sep 30 14:06:10 crc kubenswrapper[4840]: I0930 14:06:10.551356 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-sszwr" event={"ID":"18d21186-8397-400e-a9df-5e4e41e6739e","Type":"ContainerStarted","Data":"ae7bf5fcd0d76d9b6126c1bb5e75b8e757bc77bd2eb5bc6e25bd23b96e2f72b3"} Sep 30 14:06:16 crc kubenswrapper[4840]: I0930 14:06:16.587700 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-f4kpb" event={"ID":"8df1331d-54ff-47cd-80c9-c18a5ab2df53","Type":"ContainerStarted","Data":"02b313c2bd23e696f71f62c92e5157ced083a0666ee163d16d226580546e07d0"} Sep 30 14:06:16 crc kubenswrapper[4840]: I0930 14:06:16.589446 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-sszwr" event={"ID":"18d21186-8397-400e-a9df-5e4e41e6739e","Type":"ContainerStarted","Data":"5151c4b959c047c2f764666ba64c1b161c861c95a1a33ae1d37a7167c3b6cf2b"} Sep 30 14:06:16 crc kubenswrapper[4840]: I0930 14:06:16.589746 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-sszwr" Sep 30 14:06:16 crc kubenswrapper[4840]: I0930 14:06:16.603156 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-f4kpb" podStartSLOduration=1.712032959 podStartE2EDuration="7.603133511s" podCreationTimestamp="2025-09-30 14:06:09 +0000 UTC" firstStartedPulling="2025-09-30 14:06:09.68141038 +0000 UTC m=+598.310496803" lastFinishedPulling="2025-09-30 14:06:15.572510932 +0000 UTC m=+604.201597355" observedRunningTime="2025-09-30 14:06:16.602016042 +0000 UTC m=+605.231102455" watchObservedRunningTime="2025-09-30 14:06:16.603133511 +0000 UTC m=+605.232219934" Sep 30 14:06:16 crc kubenswrapper[4840]: I0930 14:06:16.618181 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-sszwr" podStartSLOduration=1.761616798 podStartE2EDuration="7.618157772s" podCreationTimestamp="2025-09-30 14:06:09 +0000 UTC" firstStartedPulling="2025-09-30 14:06:09.716065531 +0000 UTC m=+598.345151954" lastFinishedPulling="2025-09-30 14:06:15.572606495 +0000 UTC m=+604.201692928" observedRunningTime="2025-09-30 14:06:16.614603122 +0000 UTC m=+605.243689565" watchObservedRunningTime="2025-09-30 14:06:16.618157772 +0000 UTC m=+605.247244195" Sep 30 14:06:17 crc kubenswrapper[4840]: I0930 14:06:17.597037 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-hjfs2" event={"ID":"11b92f8e-b36f-47e5-8d1a-d3c83ccbd870","Type":"ContainerStarted","Data":"cd29c69d097aea98d48b5f20a78a8bcba8c016088470cc671941cb2fbfad97e5"} Sep 30 14:06:17 crc kubenswrapper[4840]: I0930 14:06:17.618815 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-hjfs2" podStartSLOduration=1.869402918 podStartE2EDuration="8.618797239s" podCreationTimestamp="2025-09-30 14:06:09 +0000 UTC" firstStartedPulling="2025-09-30 14:06:09.729955444 +0000 UTC m=+598.359041867" lastFinishedPulling="2025-09-30 14:06:16.479349765 +0000 UTC m=+605.108436188" observedRunningTime="2025-09-30 14:06:17.616175552 +0000 UTC m=+606.245261995" watchObservedRunningTime="2025-09-30 14:06:17.618797239 +0000 UTC m=+606.247883662" Sep 30 14:06:19 crc kubenswrapper[4840]: I0930 14:06:19.187077 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-kfx69"] Sep 30 14:06:19 crc kubenswrapper[4840]: I0930 14:06:19.187741 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="ovn-controller" containerID="cri-o://7a364a13c773fe00ee4e022a286821013d3a681acc16567e66b1c33c7e18ee3a" gracePeriod=30 Sep 30 14:06:19 crc kubenswrapper[4840]: I0930 14:06:19.187886 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="ovn-acl-logging" containerID="cri-o://b9549ce85242f1f19b3a52e603caec58bb4c3e2449eb50d8c4525f56dfc93e98" gracePeriod=30 Sep 30 14:06:19 crc kubenswrapper[4840]: I0930 14:06:19.187856 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="northd" containerID="cri-o://bc4e5d3625aa2d5f94307cedddf72cc48f17f327a044501144e35422d670d555" gracePeriod=30 Sep 30 14:06:19 crc kubenswrapper[4840]: I0930 14:06:19.187904 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="kube-rbac-proxy-node" containerID="cri-o://f84052a424fd6279d70b54902ebdcf430c8aa53b02406b52fc261db9d4d88ad5" gracePeriod=30 Sep 30 14:06:19 crc kubenswrapper[4840]: I0930 14:06:19.188065 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="nbdb" containerID="cri-o://d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134" gracePeriod=30 Sep 30 14:06:19 crc kubenswrapper[4840]: I0930 14:06:19.188074 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://f27a94b18137fa991459484406b1c5bd2b2dbb52f527d5f13156ba8e5b8d0146" gracePeriod=30 Sep 30 14:06:19 crc kubenswrapper[4840]: I0930 14:06:19.188110 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="sbdb" containerID="cri-o://c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d" gracePeriod=30 Sep 30 14:06:19 crc kubenswrapper[4840]: I0930 14:06:19.211997 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="ovnkube-controller" containerID="cri-o://6fdd66531df81c022d7053dc10bb0f7c09da2dabfacf7993d7439a395f2f453b" gracePeriod=30 Sep 30 14:06:19 crc kubenswrapper[4840]: E0930 14:06:19.403687 4840 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134 is running failed: container process not found" containerID="d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Sep 30 14:06:19 crc kubenswrapper[4840]: E0930 14:06:19.403905 4840 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d is running failed: container process not found" containerID="c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Sep 30 14:06:19 crc kubenswrapper[4840]: E0930 14:06:19.404317 4840 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134 is running failed: container process not found" containerID="d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Sep 30 14:06:19 crc kubenswrapper[4840]: E0930 14:06:19.404447 4840 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d is running failed: container process not found" containerID="c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Sep 30 14:06:19 crc kubenswrapper[4840]: E0930 14:06:19.404880 4840 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d is running failed: container process not found" containerID="c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Sep 30 14:06:19 crc kubenswrapper[4840]: E0930 14:06:19.404889 4840 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134 is running failed: container process not found" containerID="d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Sep 30 14:06:19 crc kubenswrapper[4840]: E0930 14:06:19.404918 4840 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d is running failed: container process not found" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="sbdb" Sep 30 14:06:19 crc kubenswrapper[4840]: E0930 14:06:19.404927 4840 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134 is running failed: container process not found" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="nbdb" Sep 30 14:06:19 crc kubenswrapper[4840]: I0930 14:06:19.609115 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kfx69_2ff48c28-d076-46e8-a93f-9630989f81e8/ovnkube-controller/3.log" Sep 30 14:06:19 crc kubenswrapper[4840]: I0930 14:06:19.611988 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kfx69_2ff48c28-d076-46e8-a93f-9630989f81e8/ovn-acl-logging/0.log" Sep 30 14:06:19 crc kubenswrapper[4840]: I0930 14:06:19.612506 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kfx69_2ff48c28-d076-46e8-a93f-9630989f81e8/ovn-controller/0.log" Sep 30 14:06:19 crc kubenswrapper[4840]: I0930 14:06:19.612931 4840 generic.go:334] "Generic (PLEG): container finished" podID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerID="6fdd66531df81c022d7053dc10bb0f7c09da2dabfacf7993d7439a395f2f453b" exitCode=0 Sep 30 14:06:19 crc kubenswrapper[4840]: I0930 14:06:19.612959 4840 generic.go:334] "Generic (PLEG): container finished" podID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerID="c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d" exitCode=0 Sep 30 14:06:19 crc kubenswrapper[4840]: I0930 14:06:19.612969 4840 generic.go:334] "Generic (PLEG): container finished" podID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerID="d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134" exitCode=0 Sep 30 14:06:19 crc kubenswrapper[4840]: I0930 14:06:19.612981 4840 generic.go:334] "Generic (PLEG): container finished" podID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerID="bc4e5d3625aa2d5f94307cedddf72cc48f17f327a044501144e35422d670d555" exitCode=0 Sep 30 14:06:19 crc kubenswrapper[4840]: I0930 14:06:19.612989 4840 generic.go:334] "Generic (PLEG): container finished" podID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerID="f27a94b18137fa991459484406b1c5bd2b2dbb52f527d5f13156ba8e5b8d0146" exitCode=0 Sep 30 14:06:19 crc kubenswrapper[4840]: I0930 14:06:19.612997 4840 generic.go:334] "Generic (PLEG): container finished" podID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerID="f84052a424fd6279d70b54902ebdcf430c8aa53b02406b52fc261db9d4d88ad5" exitCode=0 Sep 30 14:06:19 crc kubenswrapper[4840]: I0930 14:06:19.613006 4840 generic.go:334] "Generic (PLEG): container finished" podID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerID="b9549ce85242f1f19b3a52e603caec58bb4c3e2449eb50d8c4525f56dfc93e98" exitCode=143 Sep 30 14:06:19 crc kubenswrapper[4840]: I0930 14:06:19.613016 4840 generic.go:334] "Generic (PLEG): container finished" podID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerID="7a364a13c773fe00ee4e022a286821013d3a681acc16567e66b1c33c7e18ee3a" exitCode=143 Sep 30 14:06:19 crc kubenswrapper[4840]: I0930 14:06:19.613011 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" event={"ID":"2ff48c28-d076-46e8-a93f-9630989f81e8","Type":"ContainerDied","Data":"6fdd66531df81c022d7053dc10bb0f7c09da2dabfacf7993d7439a395f2f453b"} Sep 30 14:06:19 crc kubenswrapper[4840]: I0930 14:06:19.613057 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" event={"ID":"2ff48c28-d076-46e8-a93f-9630989f81e8","Type":"ContainerDied","Data":"c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d"} Sep 30 14:06:19 crc kubenswrapper[4840]: I0930 14:06:19.613069 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" event={"ID":"2ff48c28-d076-46e8-a93f-9630989f81e8","Type":"ContainerDied","Data":"d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134"} Sep 30 14:06:19 crc kubenswrapper[4840]: I0930 14:06:19.613080 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" event={"ID":"2ff48c28-d076-46e8-a93f-9630989f81e8","Type":"ContainerDied","Data":"bc4e5d3625aa2d5f94307cedddf72cc48f17f327a044501144e35422d670d555"} Sep 30 14:06:19 crc kubenswrapper[4840]: I0930 14:06:19.613092 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" event={"ID":"2ff48c28-d076-46e8-a93f-9630989f81e8","Type":"ContainerDied","Data":"f27a94b18137fa991459484406b1c5bd2b2dbb52f527d5f13156ba8e5b8d0146"} Sep 30 14:06:19 crc kubenswrapper[4840]: I0930 14:06:19.613101 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" event={"ID":"2ff48c28-d076-46e8-a93f-9630989f81e8","Type":"ContainerDied","Data":"f84052a424fd6279d70b54902ebdcf430c8aa53b02406b52fc261db9d4d88ad5"} Sep 30 14:06:19 crc kubenswrapper[4840]: I0930 14:06:19.613111 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" event={"ID":"2ff48c28-d076-46e8-a93f-9630989f81e8","Type":"ContainerDied","Data":"b9549ce85242f1f19b3a52e603caec58bb4c3e2449eb50d8c4525f56dfc93e98"} Sep 30 14:06:19 crc kubenswrapper[4840]: I0930 14:06:19.613116 4840 scope.go:117] "RemoveContainer" containerID="109628b3a230ce972e436d712479c6f8071e34a983f83d79555ffea902ca2e55" Sep 30 14:06:19 crc kubenswrapper[4840]: I0930 14:06:19.613122 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" event={"ID":"2ff48c28-d076-46e8-a93f-9630989f81e8","Type":"ContainerDied","Data":"7a364a13c773fe00ee4e022a286821013d3a681acc16567e66b1c33c7e18ee3a"} Sep 30 14:06:19 crc kubenswrapper[4840]: I0930 14:06:19.614782 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bwvl2_cbe233ee-1ea7-433e-a53a-e4a668f739ee/kube-multus/2.log" Sep 30 14:06:19 crc kubenswrapper[4840]: I0930 14:06:19.615176 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bwvl2_cbe233ee-1ea7-433e-a53a-e4a668f739ee/kube-multus/1.log" Sep 30 14:06:19 crc kubenswrapper[4840]: I0930 14:06:19.615213 4840 generic.go:334] "Generic (PLEG): container finished" podID="cbe233ee-1ea7-433e-a53a-e4a668f739ee" containerID="16d3d0c48c4c35832f75f43479de2114774ae32ec74b4f01cc9e8a63dee564c6" exitCode=2 Sep 30 14:06:19 crc kubenswrapper[4840]: I0930 14:06:19.615232 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bwvl2" event={"ID":"cbe233ee-1ea7-433e-a53a-e4a668f739ee","Type":"ContainerDied","Data":"16d3d0c48c4c35832f75f43479de2114774ae32ec74b4f01cc9e8a63dee564c6"} Sep 30 14:06:19 crc kubenswrapper[4840]: I0930 14:06:19.615644 4840 scope.go:117] "RemoveContainer" containerID="16d3d0c48c4c35832f75f43479de2114774ae32ec74b4f01cc9e8a63dee564c6" Sep 30 14:06:19 crc kubenswrapper[4840]: E0930 14:06:19.615868 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-bwvl2_openshift-multus(cbe233ee-1ea7-433e-a53a-e4a668f739ee)\"" pod="openshift-multus/multus-bwvl2" podUID="cbe233ee-1ea7-433e-a53a-e4a668f739ee" Sep 30 14:06:19 crc kubenswrapper[4840]: I0930 14:06:19.661170 4840 scope.go:117] "RemoveContainer" containerID="a3eb5795ae35b00eba9462e95fdba89c2dcf0442843c32ec05edd710dfe507fe" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.011421 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kfx69_2ff48c28-d076-46e8-a93f-9630989f81e8/ovn-acl-logging/0.log" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.011991 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kfx69_2ff48c28-d076-46e8-a93f-9630989f81e8/ovn-controller/0.log" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.012508 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.065937 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-sdwjl"] Sep 30 14:06:20 crc kubenswrapper[4840]: E0930 14:06:20.066201 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="ovn-acl-logging" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.066218 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="ovn-acl-logging" Sep 30 14:06:20 crc kubenswrapper[4840]: E0930 14:06:20.066232 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="kube-rbac-proxy-node" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.066241 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="kube-rbac-proxy-node" Sep 30 14:06:20 crc kubenswrapper[4840]: E0930 14:06:20.066252 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="kubecfg-setup" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.066260 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="kubecfg-setup" Sep 30 14:06:20 crc kubenswrapper[4840]: E0930 14:06:20.066274 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="northd" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.066282 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="northd" Sep 30 14:06:20 crc kubenswrapper[4840]: E0930 14:06:20.066293 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="ovnkube-controller" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.066300 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="ovnkube-controller" Sep 30 14:06:20 crc kubenswrapper[4840]: E0930 14:06:20.066313 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="ovnkube-controller" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.066320 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="ovnkube-controller" Sep 30 14:06:20 crc kubenswrapper[4840]: E0930 14:06:20.066330 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="sbdb" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.066338 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="sbdb" Sep 30 14:06:20 crc kubenswrapper[4840]: E0930 14:06:20.066348 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="nbdb" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.066355 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="nbdb" Sep 30 14:06:20 crc kubenswrapper[4840]: E0930 14:06:20.066365 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="kube-rbac-proxy-ovn-metrics" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.066375 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="kube-rbac-proxy-ovn-metrics" Sep 30 14:06:20 crc kubenswrapper[4840]: E0930 14:06:20.066389 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="ovn-controller" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.066398 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="ovn-controller" Sep 30 14:06:20 crc kubenswrapper[4840]: E0930 14:06:20.066410 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="ovnkube-controller" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.066417 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="ovnkube-controller" Sep 30 14:06:20 crc kubenswrapper[4840]: E0930 14:06:20.066426 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="ovnkube-controller" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.066434 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="ovnkube-controller" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.066577 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="sbdb" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.066594 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="ovnkube-controller" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.066603 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="ovnkube-controller" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.066611 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="northd" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.066619 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="ovn-controller" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.066629 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="ovn-acl-logging" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.066642 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="ovnkube-controller" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.066652 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="ovnkube-controller" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.066660 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="kube-rbac-proxy-ovn-metrics" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.066671 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="nbdb" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.066681 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="kube-rbac-proxy-node" Sep 30 14:06:20 crc kubenswrapper[4840]: E0930 14:06:20.066810 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="ovnkube-controller" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.066822 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="ovnkube-controller" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.066965 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" containerName="ovnkube-controller" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.068525 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.142974 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-systemd-units\") pod \"2ff48c28-d076-46e8-a93f-9630989f81e8\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.143010 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-run-netns\") pod \"2ff48c28-d076-46e8-a93f-9630989f81e8\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.143062 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2ff48c28-d076-46e8-a93f-9630989f81e8-env-overrides\") pod \"2ff48c28-d076-46e8-a93f-9630989f81e8\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.143081 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-slash\") pod \"2ff48c28-d076-46e8-a93f-9630989f81e8\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.143088 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "2ff48c28-d076-46e8-a93f-9630989f81e8" (UID: "2ff48c28-d076-46e8-a93f-9630989f81e8"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.143103 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-run-ovn-kubernetes\") pod \"2ff48c28-d076-46e8-a93f-9630989f81e8\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.143121 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2ff48c28-d076-46e8-a93f-9630989f81e8-ovn-node-metrics-cert\") pod \"2ff48c28-d076-46e8-a93f-9630989f81e8\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.143150 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-run-openvswitch\") pod \"2ff48c28-d076-46e8-a93f-9630989f81e8\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.143171 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2ff48c28-d076-46e8-a93f-9630989f81e8-ovnkube-config\") pod \"2ff48c28-d076-46e8-a93f-9630989f81e8\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.143186 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-kubelet\") pod \"2ff48c28-d076-46e8-a93f-9630989f81e8\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.143207 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tbvzp\" (UniqueName: \"kubernetes.io/projected/2ff48c28-d076-46e8-a93f-9630989f81e8-kube-api-access-tbvzp\") pod \"2ff48c28-d076-46e8-a93f-9630989f81e8\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.143187 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "2ff48c28-d076-46e8-a93f-9630989f81e8" (UID: "2ff48c28-d076-46e8-a93f-9630989f81e8"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.143210 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "2ff48c28-d076-46e8-a93f-9630989f81e8" (UID: "2ff48c28-d076-46e8-a93f-9630989f81e8"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.143210 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-slash" (OuterVolumeSpecName: "host-slash") pod "2ff48c28-d076-46e8-a93f-9630989f81e8" (UID: "2ff48c28-d076-46e8-a93f-9630989f81e8"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.143225 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2ff48c28-d076-46e8-a93f-9630989f81e8-ovnkube-script-lib\") pod \"2ff48c28-d076-46e8-a93f-9630989f81e8\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.143378 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-var-lib-openvswitch\") pod \"2ff48c28-d076-46e8-a93f-9630989f81e8\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.143420 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-log-socket\") pod \"2ff48c28-d076-46e8-a93f-9630989f81e8\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.143432 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ff48c28-d076-46e8-a93f-9630989f81e8-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "2ff48c28-d076-46e8-a93f-9630989f81e8" (UID: "2ff48c28-d076-46e8-a93f-9630989f81e8"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.143453 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "2ff48c28-d076-46e8-a93f-9630989f81e8" (UID: "2ff48c28-d076-46e8-a93f-9630989f81e8"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.143479 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-etc-openvswitch\") pod \"2ff48c28-d076-46e8-a93f-9630989f81e8\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.143513 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-cni-bin\") pod \"2ff48c28-d076-46e8-a93f-9630989f81e8\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.143566 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-node-log\") pod \"2ff48c28-d076-46e8-a93f-9630989f81e8\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.143600 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-run-systemd\") pod \"2ff48c28-d076-46e8-a93f-9630989f81e8\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.143612 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "2ff48c28-d076-46e8-a93f-9630989f81e8" (UID: "2ff48c28-d076-46e8-a93f-9630989f81e8"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.143624 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ff48c28-d076-46e8-a93f-9630989f81e8-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "2ff48c28-d076-46e8-a93f-9630989f81e8" (UID: "2ff48c28-d076-46e8-a93f-9630989f81e8"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.143631 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"2ff48c28-d076-46e8-a93f-9630989f81e8\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.143642 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "2ff48c28-d076-46e8-a93f-9630989f81e8" (UID: "2ff48c28-d076-46e8-a93f-9630989f81e8"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.143661 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "2ff48c28-d076-46e8-a93f-9630989f81e8" (UID: "2ff48c28-d076-46e8-a93f-9630989f81e8"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.143668 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-log-socket" (OuterVolumeSpecName: "log-socket") pod "2ff48c28-d076-46e8-a93f-9630989f81e8" (UID: "2ff48c28-d076-46e8-a93f-9630989f81e8"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.143697 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "2ff48c28-d076-46e8-a93f-9630989f81e8" (UID: "2ff48c28-d076-46e8-a93f-9630989f81e8"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.143696 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ff48c28-d076-46e8-a93f-9630989f81e8-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "2ff48c28-d076-46e8-a93f-9630989f81e8" (UID: "2ff48c28-d076-46e8-a93f-9630989f81e8"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.143702 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-node-log" (OuterVolumeSpecName: "node-log") pod "2ff48c28-d076-46e8-a93f-9630989f81e8" (UID: "2ff48c28-d076-46e8-a93f-9630989f81e8"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.143729 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "2ff48c28-d076-46e8-a93f-9630989f81e8" (UID: "2ff48c28-d076-46e8-a93f-9630989f81e8"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.143741 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "2ff48c28-d076-46e8-a93f-9630989f81e8" (UID: "2ff48c28-d076-46e8-a93f-9630989f81e8"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.143711 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-run-ovn\") pod \"2ff48c28-d076-46e8-a93f-9630989f81e8\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.143801 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-cni-netd\") pod \"2ff48c28-d076-46e8-a93f-9630989f81e8\" (UID: \"2ff48c28-d076-46e8-a93f-9630989f81e8\") " Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.143898 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "2ff48c28-d076-46e8-a93f-9630989f81e8" (UID: "2ff48c28-d076-46e8-a93f-9630989f81e8"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.144034 4840 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.144049 4840 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-run-ovn\") on node \"crc\" DevicePath \"\"" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.144058 4840 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-cni-netd\") on node \"crc\" DevicePath \"\"" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.144067 4840 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-systemd-units\") on node \"crc\" DevicePath \"\"" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.144077 4840 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-run-netns\") on node \"crc\" DevicePath \"\"" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.144087 4840 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2ff48c28-d076-46e8-a93f-9630989f81e8-env-overrides\") on node \"crc\" DevicePath \"\"" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.144096 4840 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-slash\") on node \"crc\" DevicePath \"\"" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.144106 4840 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.144116 4840 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-run-openvswitch\") on node \"crc\" DevicePath \"\"" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.144127 4840 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2ff48c28-d076-46e8-a93f-9630989f81e8-ovnkube-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.144137 4840 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-kubelet\") on node \"crc\" DevicePath \"\"" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.144147 4840 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2ff48c28-d076-46e8-a93f-9630989f81e8-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.144157 4840 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.144167 4840 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-log-socket\") on node \"crc\" DevicePath \"\"" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.144177 4840 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.144186 4840 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-host-cni-bin\") on node \"crc\" DevicePath \"\"" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.144195 4840 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-node-log\") on node \"crc\" DevicePath \"\"" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.151202 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ff48c28-d076-46e8-a93f-9630989f81e8-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "2ff48c28-d076-46e8-a93f-9630989f81e8" (UID: "2ff48c28-d076-46e8-a93f-9630989f81e8"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.151532 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ff48c28-d076-46e8-a93f-9630989f81e8-kube-api-access-tbvzp" (OuterVolumeSpecName: "kube-api-access-tbvzp") pod "2ff48c28-d076-46e8-a93f-9630989f81e8" (UID: "2ff48c28-d076-46e8-a93f-9630989f81e8"). InnerVolumeSpecName "kube-api-access-tbvzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.160423 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "2ff48c28-d076-46e8-a93f-9630989f81e8" (UID: "2ff48c28-d076-46e8-a93f-9630989f81e8"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.244983 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjfb6\" (UniqueName: \"kubernetes.io/projected/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-kube-api-access-pjfb6\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.245058 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-systemd-units\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.245088 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-host-kubelet\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.245116 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-ovn-node-metrics-cert\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.245140 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-ovnkube-config\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.245162 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-run-systemd\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.245379 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-run-openvswitch\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.245443 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-var-lib-openvswitch\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.245485 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-node-log\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.245515 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-host-slash\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.245643 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-run-ovn\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.245689 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-host-run-netns\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.245733 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-ovnkube-script-lib\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.245812 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-host-cni-bin\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.245844 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-etc-openvswitch\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.245889 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-env-overrides\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.245925 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-log-socket\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.245962 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.245987 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-host-cni-netd\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.246135 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-host-run-ovn-kubernetes\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.246365 4840 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2ff48c28-d076-46e8-a93f-9630989f81e8-run-systemd\") on node \"crc\" DevicePath \"\"" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.246394 4840 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2ff48c28-d076-46e8-a93f-9630989f81e8-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.246413 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tbvzp\" (UniqueName: \"kubernetes.io/projected/2ff48c28-d076-46e8-a93f-9630989f81e8-kube-api-access-tbvzp\") on node \"crc\" DevicePath \"\"" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.347670 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-ovn-node-metrics-cert\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.347729 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-run-systemd\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.347748 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-ovnkube-config\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.347789 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-run-openvswitch\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.347813 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-var-lib-openvswitch\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.347820 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-run-systemd\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.347862 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-node-log\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.347834 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-node-log\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.347894 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-run-openvswitch\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.347905 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-host-slash\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.347920 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-var-lib-openvswitch\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.347942 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-run-ovn\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.347982 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-host-run-netns\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.348019 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-ovnkube-script-lib\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.348055 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-host-cni-bin\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.348083 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-host-run-netns\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.348086 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-etc-openvswitch\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.348111 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-host-cni-bin\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.348062 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-run-ovn\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.348122 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-env-overrides\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.348155 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-log-socket\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.348190 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-host-cni-netd\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.348227 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.348267 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-host-run-ovn-kubernetes\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.348191 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-log-socket\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.348144 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-host-slash\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.348353 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-host-cni-netd\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.348421 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-host-run-ovn-kubernetes\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.348286 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-etc-openvswitch\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.348450 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.348518 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjfb6\" (UniqueName: \"kubernetes.io/projected/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-kube-api-access-pjfb6\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.348575 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-systemd-units\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.348607 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-host-kubelet\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.348718 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-systemd-units\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.348806 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-env-overrides\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.348836 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-host-kubelet\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.349251 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-ovnkube-config\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.350207 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-ovnkube-script-lib\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.352027 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-ovn-node-metrics-cert\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.370619 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjfb6\" (UniqueName: \"kubernetes.io/projected/c40ad6dd-97d8-4ec0-8b04-84ac7f25c969-kube-api-access-pjfb6\") pod \"ovnkube-node-sdwjl\" (UID: \"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969\") " pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.390489 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:20 crc kubenswrapper[4840]: W0930 14:06:20.412122 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc40ad6dd_97d8_4ec0_8b04_84ac7f25c969.slice/crio-e4d1c10cf81ef8905c4918b1dcae303a4b2975b6ac7beca8b4b371c0bcf9b7ee WatchSource:0}: Error finding container e4d1c10cf81ef8905c4918b1dcae303a4b2975b6ac7beca8b4b371c0bcf9b7ee: Status 404 returned error can't find the container with id e4d1c10cf81ef8905c4918b1dcae303a4b2975b6ac7beca8b4b371c0bcf9b7ee Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.622406 4840 generic.go:334] "Generic (PLEG): container finished" podID="c40ad6dd-97d8-4ec0-8b04-84ac7f25c969" containerID="46700d724321e0c87dbc04ebf5ca10884ec4d26beeb3a16a5a46e9f32845c2dd" exitCode=0 Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.622443 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" event={"ID":"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969","Type":"ContainerDied","Data":"46700d724321e0c87dbc04ebf5ca10884ec4d26beeb3a16a5a46e9f32845c2dd"} Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.622796 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" event={"ID":"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969","Type":"ContainerStarted","Data":"e4d1c10cf81ef8905c4918b1dcae303a4b2975b6ac7beca8b4b371c0bcf9b7ee"} Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.629852 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kfx69_2ff48c28-d076-46e8-a93f-9630989f81e8/ovn-acl-logging/0.log" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.630524 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kfx69_2ff48c28-d076-46e8-a93f-9630989f81e8/ovn-controller/0.log" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.630970 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" event={"ID":"2ff48c28-d076-46e8-a93f-9630989f81e8","Type":"ContainerDied","Data":"a9718c1e08d1bd053584576c34795f618d2cd2f04bffabc0aeb0c96f577efab8"} Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.631007 4840 scope.go:117] "RemoveContainer" containerID="6fdd66531df81c022d7053dc10bb0f7c09da2dabfacf7993d7439a395f2f453b" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.631050 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-kfx69" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.634269 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bwvl2_cbe233ee-1ea7-433e-a53a-e4a668f739ee/kube-multus/2.log" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.650863 4840 scope.go:117] "RemoveContainer" containerID="c62da556c71d99d62440e9d114397bf28a19131362b13c4a55f1388728f7453d" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.666334 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-kfx69"] Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.669374 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-kfx69"] Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.672534 4840 scope.go:117] "RemoveContainer" containerID="d34c1e58c31c986f7e694716b0051f9022c57f94866a72f7cdfe0b9c4beba134" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.684182 4840 scope.go:117] "RemoveContainer" containerID="bc4e5d3625aa2d5f94307cedddf72cc48f17f327a044501144e35422d670d555" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.697494 4840 scope.go:117] "RemoveContainer" containerID="f27a94b18137fa991459484406b1c5bd2b2dbb52f527d5f13156ba8e5b8d0146" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.710109 4840 scope.go:117] "RemoveContainer" containerID="f84052a424fd6279d70b54902ebdcf430c8aa53b02406b52fc261db9d4d88ad5" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.723304 4840 scope.go:117] "RemoveContainer" containerID="b9549ce85242f1f19b3a52e603caec58bb4c3e2449eb50d8c4525f56dfc93e98" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.737616 4840 scope.go:117] "RemoveContainer" containerID="7a364a13c773fe00ee4e022a286821013d3a681acc16567e66b1c33c7e18ee3a" Sep 30 14:06:20 crc kubenswrapper[4840]: I0930 14:06:20.751853 4840 scope.go:117] "RemoveContainer" containerID="c192647257664f846ed760df03302a55ff5620b4ce8cba09fa50d21bbbbf5f30" Sep 30 14:06:21 crc kubenswrapper[4840]: I0930 14:06:21.645703 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" event={"ID":"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969","Type":"ContainerStarted","Data":"af1facf1e4dc2e409c478a3d88cdbd98ff97c377b96f054ddcc43b311c8c930d"} Sep 30 14:06:21 crc kubenswrapper[4840]: I0930 14:06:21.646072 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" event={"ID":"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969","Type":"ContainerStarted","Data":"212819e50005975e075a2683ccb7a93440a6871a118b2df576c398566830d14f"} Sep 30 14:06:21 crc kubenswrapper[4840]: I0930 14:06:21.646090 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" event={"ID":"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969","Type":"ContainerStarted","Data":"f0863b49837a966aaa59d126e47b5139b4bd98548233fca3aa5fe75b3f7209e8"} Sep 30 14:06:21 crc kubenswrapper[4840]: I0930 14:06:21.646102 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" event={"ID":"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969","Type":"ContainerStarted","Data":"6e04ab63cf2931e2293c28d414a30749584d6b90f0049809b9c8b33ad59d03a0"} Sep 30 14:06:21 crc kubenswrapper[4840]: I0930 14:06:21.646114 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" event={"ID":"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969","Type":"ContainerStarted","Data":"6363e0335897990b7e393036842daf12e1c522e4943b2419d87242135c461c89"} Sep 30 14:06:21 crc kubenswrapper[4840]: I0930 14:06:21.646126 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" event={"ID":"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969","Type":"ContainerStarted","Data":"c9d3636b9958b1e803022feed59a1d5ccbcf12200bfffd166864abfd724dea79"} Sep 30 14:06:21 crc kubenswrapper[4840]: I0930 14:06:21.871980 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:06:21 crc kubenswrapper[4840]: I0930 14:06:21.872044 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:06:22 crc kubenswrapper[4840]: I0930 14:06:22.123719 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ff48c28-d076-46e8-a93f-9630989f81e8" path="/var/lib/kubelet/pods/2ff48c28-d076-46e8-a93f-9630989f81e8/volumes" Sep 30 14:06:23 crc kubenswrapper[4840]: I0930 14:06:23.657883 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" event={"ID":"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969","Type":"ContainerStarted","Data":"5ea79bfd6ee9475087cb41698cb6ade0e6b24cdf718b110a90a848cdcac00e47"} Sep 30 14:06:24 crc kubenswrapper[4840]: I0930 14:06:24.453783 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-sszwr" Sep 30 14:06:26 crc kubenswrapper[4840]: I0930 14:06:26.673913 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" event={"ID":"c40ad6dd-97d8-4ec0-8b04-84ac7f25c969","Type":"ContainerStarted","Data":"4d32ec3b82074be9a835a407a34c5972be4240f96d4d11472dba4c583c6751fa"} Sep 30 14:06:26 crc kubenswrapper[4840]: I0930 14:06:26.674423 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:26 crc kubenswrapper[4840]: I0930 14:06:26.674434 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:26 crc kubenswrapper[4840]: I0930 14:06:26.674443 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:26 crc kubenswrapper[4840]: I0930 14:06:26.701106 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" podStartSLOduration=6.701085568 podStartE2EDuration="6.701085568s" podCreationTimestamp="2025-09-30 14:06:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:06:26.699527449 +0000 UTC m=+615.328613902" watchObservedRunningTime="2025-09-30 14:06:26.701085568 +0000 UTC m=+615.330172001" Sep 30 14:06:26 crc kubenswrapper[4840]: I0930 14:06:26.706785 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:26 crc kubenswrapper[4840]: I0930 14:06:26.707511 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:34 crc kubenswrapper[4840]: I0930 14:06:34.116397 4840 scope.go:117] "RemoveContainer" containerID="16d3d0c48c4c35832f75f43479de2114774ae32ec74b4f01cc9e8a63dee564c6" Sep 30 14:06:34 crc kubenswrapper[4840]: E0930 14:06:34.117185 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-bwvl2_openshift-multus(cbe233ee-1ea7-433e-a53a-e4a668f739ee)\"" pod="openshift-multus/multus-bwvl2" podUID="cbe233ee-1ea7-433e-a53a-e4a668f739ee" Sep 30 14:06:47 crc kubenswrapper[4840]: I0930 14:06:47.116628 4840 scope.go:117] "RemoveContainer" containerID="16d3d0c48c4c35832f75f43479de2114774ae32ec74b4f01cc9e8a63dee564c6" Sep 30 14:06:47 crc kubenswrapper[4840]: I0930 14:06:47.786351 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bwvl2_cbe233ee-1ea7-433e-a53a-e4a668f739ee/kube-multus/2.log" Sep 30 14:06:47 crc kubenswrapper[4840]: I0930 14:06:47.786725 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bwvl2" event={"ID":"cbe233ee-1ea7-433e-a53a-e4a668f739ee","Type":"ContainerStarted","Data":"f234f11cfab0f0c87b1c566ee1a902c3d71904148c333c96db6f3275adf1ea5b"} Sep 30 14:06:50 crc kubenswrapper[4840]: I0930 14:06:50.411922 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-sdwjl" Sep 30 14:06:51 crc kubenswrapper[4840]: I0930 14:06:51.872206 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:06:51 crc kubenswrapper[4840]: I0930 14:06:51.872719 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:06:51 crc kubenswrapper[4840]: I0930 14:06:51.872792 4840 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-747gk" Sep 30 14:06:51 crc kubenswrapper[4840]: I0930 14:06:51.873689 4840 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d12912f690026de70f88356c602b1439b6380a977591bd20b79ac057d9633343"} pod="openshift-machine-config-operator/machine-config-daemon-747gk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 14:06:51 crc kubenswrapper[4840]: I0930 14:06:51.873784 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" containerID="cri-o://d12912f690026de70f88356c602b1439b6380a977591bd20b79ac057d9633343" gracePeriod=600 Sep 30 14:06:52 crc kubenswrapper[4840]: I0930 14:06:52.815146 4840 generic.go:334] "Generic (PLEG): container finished" podID="10e8b890-7f20-4a36-8e03-898620cf599a" containerID="d12912f690026de70f88356c602b1439b6380a977591bd20b79ac057d9633343" exitCode=0 Sep 30 14:06:52 crc kubenswrapper[4840]: I0930 14:06:52.815286 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" event={"ID":"10e8b890-7f20-4a36-8e03-898620cf599a","Type":"ContainerDied","Data":"d12912f690026de70f88356c602b1439b6380a977591bd20b79ac057d9633343"} Sep 30 14:06:52 crc kubenswrapper[4840]: I0930 14:06:52.815838 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" event={"ID":"10e8b890-7f20-4a36-8e03-898620cf599a","Type":"ContainerStarted","Data":"999ab6436a76ee1df695f7a3673950dda45d0789af244839074f28ed47e2534b"} Sep 30 14:06:52 crc kubenswrapper[4840]: I0930 14:06:52.815890 4840 scope.go:117] "RemoveContainer" containerID="e98140cf872382a187fd61e08bfb9d9a34afdc7cbc64b4d05d64d64c86bc9322" Sep 30 14:07:09 crc kubenswrapper[4840]: I0930 14:07:09.372945 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6"] Sep 30 14:07:09 crc kubenswrapper[4840]: I0930 14:07:09.375800 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6" Sep 30 14:07:09 crc kubenswrapper[4840]: I0930 14:07:09.378768 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Sep 30 14:07:09 crc kubenswrapper[4840]: I0930 14:07:09.385181 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6"] Sep 30 14:07:09 crc kubenswrapper[4840]: I0930 14:07:09.562499 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nx5n7\" (UniqueName: \"kubernetes.io/projected/0baa1415-f354-45dd-a352-140902beba82-kube-api-access-nx5n7\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6\" (UID: \"0baa1415-f354-45dd-a352-140902beba82\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6" Sep 30 14:07:09 crc kubenswrapper[4840]: I0930 14:07:09.563076 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0baa1415-f354-45dd-a352-140902beba82-bundle\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6\" (UID: \"0baa1415-f354-45dd-a352-140902beba82\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6" Sep 30 14:07:09 crc kubenswrapper[4840]: I0930 14:07:09.563204 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0baa1415-f354-45dd-a352-140902beba82-util\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6\" (UID: \"0baa1415-f354-45dd-a352-140902beba82\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6" Sep 30 14:07:09 crc kubenswrapper[4840]: I0930 14:07:09.664212 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nx5n7\" (UniqueName: \"kubernetes.io/projected/0baa1415-f354-45dd-a352-140902beba82-kube-api-access-nx5n7\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6\" (UID: \"0baa1415-f354-45dd-a352-140902beba82\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6" Sep 30 14:07:09 crc kubenswrapper[4840]: I0930 14:07:09.664290 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0baa1415-f354-45dd-a352-140902beba82-bundle\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6\" (UID: \"0baa1415-f354-45dd-a352-140902beba82\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6" Sep 30 14:07:09 crc kubenswrapper[4840]: I0930 14:07:09.664341 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0baa1415-f354-45dd-a352-140902beba82-util\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6\" (UID: \"0baa1415-f354-45dd-a352-140902beba82\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6" Sep 30 14:07:09 crc kubenswrapper[4840]: I0930 14:07:09.664824 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0baa1415-f354-45dd-a352-140902beba82-bundle\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6\" (UID: \"0baa1415-f354-45dd-a352-140902beba82\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6" Sep 30 14:07:09 crc kubenswrapper[4840]: I0930 14:07:09.664900 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0baa1415-f354-45dd-a352-140902beba82-util\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6\" (UID: \"0baa1415-f354-45dd-a352-140902beba82\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6" Sep 30 14:07:09 crc kubenswrapper[4840]: I0930 14:07:09.686946 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nx5n7\" (UniqueName: \"kubernetes.io/projected/0baa1415-f354-45dd-a352-140902beba82-kube-api-access-nx5n7\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6\" (UID: \"0baa1415-f354-45dd-a352-140902beba82\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6" Sep 30 14:07:09 crc kubenswrapper[4840]: I0930 14:07:09.702224 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6" Sep 30 14:07:09 crc kubenswrapper[4840]: I0930 14:07:09.878363 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6"] Sep 30 14:07:09 crc kubenswrapper[4840]: I0930 14:07:09.929618 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6" event={"ID":"0baa1415-f354-45dd-a352-140902beba82","Type":"ContainerStarted","Data":"a477354ac169c75185d088748539b82f6dff1ce4ace40bc7671f5fd9c74ebb26"} Sep 30 14:07:10 crc kubenswrapper[4840]: I0930 14:07:10.936134 4840 generic.go:334] "Generic (PLEG): container finished" podID="0baa1415-f354-45dd-a352-140902beba82" containerID="d9d5391c2bbe0683ed929d919915cf3c992f452b7be76eb784f6a649dc2693c4" exitCode=0 Sep 30 14:07:10 crc kubenswrapper[4840]: I0930 14:07:10.936197 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6" event={"ID":"0baa1415-f354-45dd-a352-140902beba82","Type":"ContainerDied","Data":"d9d5391c2bbe0683ed929d919915cf3c992f452b7be76eb784f6a649dc2693c4"} Sep 30 14:07:12 crc kubenswrapper[4840]: I0930 14:07:12.950572 4840 generic.go:334] "Generic (PLEG): container finished" podID="0baa1415-f354-45dd-a352-140902beba82" containerID="dfb89f2629988a300d5c0d65635f75bfa5d7a4126396c0f0dcc785072bc01a61" exitCode=0 Sep 30 14:07:12 crc kubenswrapper[4840]: I0930 14:07:12.950754 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6" event={"ID":"0baa1415-f354-45dd-a352-140902beba82","Type":"ContainerDied","Data":"dfb89f2629988a300d5c0d65635f75bfa5d7a4126396c0f0dcc785072bc01a61"} Sep 30 14:07:13 crc kubenswrapper[4840]: I0930 14:07:13.963303 4840 generic.go:334] "Generic (PLEG): container finished" podID="0baa1415-f354-45dd-a352-140902beba82" containerID="6436a3609f25f3d626953a09f235b45b91601d3a00bdb8ca727a15104b1bce32" exitCode=0 Sep 30 14:07:13 crc kubenswrapper[4840]: I0930 14:07:13.963374 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6" event={"ID":"0baa1415-f354-45dd-a352-140902beba82","Type":"ContainerDied","Data":"6436a3609f25f3d626953a09f235b45b91601d3a00bdb8ca727a15104b1bce32"} Sep 30 14:07:15 crc kubenswrapper[4840]: I0930 14:07:15.208456 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6" Sep 30 14:07:15 crc kubenswrapper[4840]: I0930 14:07:15.331583 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0baa1415-f354-45dd-a352-140902beba82-util\") pod \"0baa1415-f354-45dd-a352-140902beba82\" (UID: \"0baa1415-f354-45dd-a352-140902beba82\") " Sep 30 14:07:15 crc kubenswrapper[4840]: I0930 14:07:15.331698 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0baa1415-f354-45dd-a352-140902beba82-bundle\") pod \"0baa1415-f354-45dd-a352-140902beba82\" (UID: \"0baa1415-f354-45dd-a352-140902beba82\") " Sep 30 14:07:15 crc kubenswrapper[4840]: I0930 14:07:15.331767 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nx5n7\" (UniqueName: \"kubernetes.io/projected/0baa1415-f354-45dd-a352-140902beba82-kube-api-access-nx5n7\") pod \"0baa1415-f354-45dd-a352-140902beba82\" (UID: \"0baa1415-f354-45dd-a352-140902beba82\") " Sep 30 14:07:15 crc kubenswrapper[4840]: I0930 14:07:15.332429 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0baa1415-f354-45dd-a352-140902beba82-bundle" (OuterVolumeSpecName: "bundle") pod "0baa1415-f354-45dd-a352-140902beba82" (UID: "0baa1415-f354-45dd-a352-140902beba82"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:07:15 crc kubenswrapper[4840]: I0930 14:07:15.336747 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0baa1415-f354-45dd-a352-140902beba82-kube-api-access-nx5n7" (OuterVolumeSpecName: "kube-api-access-nx5n7") pod "0baa1415-f354-45dd-a352-140902beba82" (UID: "0baa1415-f354-45dd-a352-140902beba82"). InnerVolumeSpecName "kube-api-access-nx5n7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:07:15 crc kubenswrapper[4840]: I0930 14:07:15.433199 4840 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0baa1415-f354-45dd-a352-140902beba82-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:07:15 crc kubenswrapper[4840]: I0930 14:07:15.433540 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nx5n7\" (UniqueName: \"kubernetes.io/projected/0baa1415-f354-45dd-a352-140902beba82-kube-api-access-nx5n7\") on node \"crc\" DevicePath \"\"" Sep 30 14:07:15 crc kubenswrapper[4840]: I0930 14:07:15.482173 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0baa1415-f354-45dd-a352-140902beba82-util" (OuterVolumeSpecName: "util") pod "0baa1415-f354-45dd-a352-140902beba82" (UID: "0baa1415-f354-45dd-a352-140902beba82"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:07:15 crc kubenswrapper[4840]: I0930 14:07:15.534221 4840 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0baa1415-f354-45dd-a352-140902beba82-util\") on node \"crc\" DevicePath \"\"" Sep 30 14:07:15 crc kubenswrapper[4840]: I0930 14:07:15.976577 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6" event={"ID":"0baa1415-f354-45dd-a352-140902beba82","Type":"ContainerDied","Data":"a477354ac169c75185d088748539b82f6dff1ce4ace40bc7671f5fd9c74ebb26"} Sep 30 14:07:15 crc kubenswrapper[4840]: I0930 14:07:15.976620 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a477354ac169c75185d088748539b82f6dff1ce4ace40bc7671f5fd9c74ebb26" Sep 30 14:07:15 crc kubenswrapper[4840]: I0930 14:07:15.976633 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6" Sep 30 14:07:17 crc kubenswrapper[4840]: I0930 14:07:17.213976 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5d6f6cfd66-wqtsv"] Sep 30 14:07:17 crc kubenswrapper[4840]: E0930 14:07:17.214203 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0baa1415-f354-45dd-a352-140902beba82" containerName="util" Sep 30 14:07:17 crc kubenswrapper[4840]: I0930 14:07:17.214215 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="0baa1415-f354-45dd-a352-140902beba82" containerName="util" Sep 30 14:07:17 crc kubenswrapper[4840]: E0930 14:07:17.214235 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0baa1415-f354-45dd-a352-140902beba82" containerName="extract" Sep 30 14:07:17 crc kubenswrapper[4840]: I0930 14:07:17.214241 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="0baa1415-f354-45dd-a352-140902beba82" containerName="extract" Sep 30 14:07:17 crc kubenswrapper[4840]: E0930 14:07:17.214250 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0baa1415-f354-45dd-a352-140902beba82" containerName="pull" Sep 30 14:07:17 crc kubenswrapper[4840]: I0930 14:07:17.214256 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="0baa1415-f354-45dd-a352-140902beba82" containerName="pull" Sep 30 14:07:17 crc kubenswrapper[4840]: I0930 14:07:17.214344 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="0baa1415-f354-45dd-a352-140902beba82" containerName="extract" Sep 30 14:07:17 crc kubenswrapper[4840]: I0930 14:07:17.214715 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-wqtsv" Sep 30 14:07:17 crc kubenswrapper[4840]: I0930 14:07:17.216992 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Sep 30 14:07:17 crc kubenswrapper[4840]: I0930 14:07:17.217172 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-w2m6f" Sep 30 14:07:17 crc kubenswrapper[4840]: I0930 14:07:17.217204 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Sep 30 14:07:17 crc kubenswrapper[4840]: I0930 14:07:17.223868 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5d6f6cfd66-wqtsv"] Sep 30 14:07:17 crc kubenswrapper[4840]: I0930 14:07:17.254504 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bb7ff\" (UniqueName: \"kubernetes.io/projected/5665a62e-e1b7-4773-afcb-009f7ffb354d-kube-api-access-bb7ff\") pod \"nmstate-operator-5d6f6cfd66-wqtsv\" (UID: \"5665a62e-e1b7-4773-afcb-009f7ffb354d\") " pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-wqtsv" Sep 30 14:07:17 crc kubenswrapper[4840]: I0930 14:07:17.355613 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bb7ff\" (UniqueName: \"kubernetes.io/projected/5665a62e-e1b7-4773-afcb-009f7ffb354d-kube-api-access-bb7ff\") pod \"nmstate-operator-5d6f6cfd66-wqtsv\" (UID: \"5665a62e-e1b7-4773-afcb-009f7ffb354d\") " pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-wqtsv" Sep 30 14:07:17 crc kubenswrapper[4840]: I0930 14:07:17.371940 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bb7ff\" (UniqueName: \"kubernetes.io/projected/5665a62e-e1b7-4773-afcb-009f7ffb354d-kube-api-access-bb7ff\") pod \"nmstate-operator-5d6f6cfd66-wqtsv\" (UID: \"5665a62e-e1b7-4773-afcb-009f7ffb354d\") " pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-wqtsv" Sep 30 14:07:17 crc kubenswrapper[4840]: I0930 14:07:17.529205 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-wqtsv" Sep 30 14:07:17 crc kubenswrapper[4840]: I0930 14:07:17.715764 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5d6f6cfd66-wqtsv"] Sep 30 14:07:17 crc kubenswrapper[4840]: W0930 14:07:17.727275 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5665a62e_e1b7_4773_afcb_009f7ffb354d.slice/crio-19e9f50cc848bc5edb2c998ba4f616f6415a5c4f0a41956a70ebdbf59ea7a8a8 WatchSource:0}: Error finding container 19e9f50cc848bc5edb2c998ba4f616f6415a5c4f0a41956a70ebdbf59ea7a8a8: Status 404 returned error can't find the container with id 19e9f50cc848bc5edb2c998ba4f616f6415a5c4f0a41956a70ebdbf59ea7a8a8 Sep 30 14:07:17 crc kubenswrapper[4840]: I0930 14:07:17.988019 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-wqtsv" event={"ID":"5665a62e-e1b7-4773-afcb-009f7ffb354d","Type":"ContainerStarted","Data":"19e9f50cc848bc5edb2c998ba4f616f6415a5c4f0a41956a70ebdbf59ea7a8a8"} Sep 30 14:07:20 crc kubenswrapper[4840]: I0930 14:07:20.006814 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-wqtsv" event={"ID":"5665a62e-e1b7-4773-afcb-009f7ffb354d","Type":"ContainerStarted","Data":"518792e523f621cf9e213a1aa29e9190f447a4188a2798ea52033a18d76f65d2"} Sep 30 14:07:20 crc kubenswrapper[4840]: I0930 14:07:20.029228 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-wqtsv" podStartSLOduration=1.037183653 podStartE2EDuration="3.029210751s" podCreationTimestamp="2025-09-30 14:07:17 +0000 UTC" firstStartedPulling="2025-09-30 14:07:17.730684915 +0000 UTC m=+666.359771328" lastFinishedPulling="2025-09-30 14:07:19.722712003 +0000 UTC m=+668.351798426" observedRunningTime="2025-09-30 14:07:20.024986934 +0000 UTC m=+668.654073397" watchObservedRunningTime="2025-09-30 14:07:20.029210751 +0000 UTC m=+668.658297194" Sep 30 14:07:20 crc kubenswrapper[4840]: I0930 14:07:20.939837 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-58fcddf996-bw56r"] Sep 30 14:07:20 crc kubenswrapper[4840]: I0930 14:07:20.941225 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58fcddf996-bw56r" Sep 30 14:07:20 crc kubenswrapper[4840]: I0930 14:07:20.946271 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-ks8m6" Sep 30 14:07:20 crc kubenswrapper[4840]: I0930 14:07:20.957635 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6d689559c5-h7fmf"] Sep 30 14:07:20 crc kubenswrapper[4840]: I0930 14:07:20.958345 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6d689559c5-h7fmf" Sep 30 14:07:20 crc kubenswrapper[4840]: I0930 14:07:20.959976 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Sep 30 14:07:20 crc kubenswrapper[4840]: I0930 14:07:20.960823 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58fcddf996-bw56r"] Sep 30 14:07:20 crc kubenswrapper[4840]: I0930 14:07:20.974590 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-gnz9p"] Sep 30 14:07:20 crc kubenswrapper[4840]: I0930 14:07:20.975226 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-gnz9p" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.003062 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/7e3e89d5-05d9-4675-9dad-83f59258f876-ovs-socket\") pod \"nmstate-handler-gnz9p\" (UID: \"7e3e89d5-05d9-4675-9dad-83f59258f876\") " pod="openshift-nmstate/nmstate-handler-gnz9p" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.003156 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4plqx\" (UniqueName: \"kubernetes.io/projected/7e3e89d5-05d9-4675-9dad-83f59258f876-kube-api-access-4plqx\") pod \"nmstate-handler-gnz9p\" (UID: \"7e3e89d5-05d9-4675-9dad-83f59258f876\") " pod="openshift-nmstate/nmstate-handler-gnz9p" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.003222 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/7e3e89d5-05d9-4675-9dad-83f59258f876-nmstate-lock\") pod \"nmstate-handler-gnz9p\" (UID: \"7e3e89d5-05d9-4675-9dad-83f59258f876\") " pod="openshift-nmstate/nmstate-handler-gnz9p" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.003256 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/7e3e89d5-05d9-4675-9dad-83f59258f876-dbus-socket\") pod \"nmstate-handler-gnz9p\" (UID: \"7e3e89d5-05d9-4675-9dad-83f59258f876\") " pod="openshift-nmstate/nmstate-handler-gnz9p" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.045040 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6d689559c5-h7fmf"] Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.073455 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-864bb6dfb5-lw82x"] Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.075488 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-lw82x" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.079967 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.080178 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.080285 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-c7vqc" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.082630 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-864bb6dfb5-lw82x"] Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.104152 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/14ccc5bb-b26a-42df-b0b9-f1ea9bcbff31-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-lw82x\" (UID: \"14ccc5bb-b26a-42df-b0b9-f1ea9bcbff31\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-lw82x" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.104202 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/7e3e89d5-05d9-4675-9dad-83f59258f876-nmstate-lock\") pod \"nmstate-handler-gnz9p\" (UID: \"7e3e89d5-05d9-4675-9dad-83f59258f876\") " pod="openshift-nmstate/nmstate-handler-gnz9p" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.104236 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/7e3e89d5-05d9-4675-9dad-83f59258f876-dbus-socket\") pod \"nmstate-handler-gnz9p\" (UID: \"7e3e89d5-05d9-4675-9dad-83f59258f876\") " pod="openshift-nmstate/nmstate-handler-gnz9p" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.104280 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-957q6\" (UniqueName: \"kubernetes.io/projected/9ed06022-f9fe-4b61-bd42-ed308118a717-kube-api-access-957q6\") pod \"nmstate-metrics-58fcddf996-bw56r\" (UID: \"9ed06022-f9fe-4b61-bd42-ed308118a717\") " pod="openshift-nmstate/nmstate-metrics-58fcddf996-bw56r" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.104328 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/7e3e89d5-05d9-4675-9dad-83f59258f876-ovs-socket\") pod \"nmstate-handler-gnz9p\" (UID: \"7e3e89d5-05d9-4675-9dad-83f59258f876\") " pod="openshift-nmstate/nmstate-handler-gnz9p" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.104381 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/e047fa81-9835-4bca-95c4-c8d24f97857b-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-h7fmf\" (UID: \"e047fa81-9835-4bca-95c4-c8d24f97857b\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-h7fmf" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.104401 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/14ccc5bb-b26a-42df-b0b9-f1ea9bcbff31-nginx-conf\") pod \"nmstate-console-plugin-864bb6dfb5-lw82x\" (UID: \"14ccc5bb-b26a-42df-b0b9-f1ea9bcbff31\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-lw82x" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.104323 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/7e3e89d5-05d9-4675-9dad-83f59258f876-nmstate-lock\") pod \"nmstate-handler-gnz9p\" (UID: \"7e3e89d5-05d9-4675-9dad-83f59258f876\") " pod="openshift-nmstate/nmstate-handler-gnz9p" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.104425 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9gzx\" (UniqueName: \"kubernetes.io/projected/14ccc5bb-b26a-42df-b0b9-f1ea9bcbff31-kube-api-access-v9gzx\") pod \"nmstate-console-plugin-864bb6dfb5-lw82x\" (UID: \"14ccc5bb-b26a-42df-b0b9-f1ea9bcbff31\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-lw82x" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.104396 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/7e3e89d5-05d9-4675-9dad-83f59258f876-ovs-socket\") pod \"nmstate-handler-gnz9p\" (UID: \"7e3e89d5-05d9-4675-9dad-83f59258f876\") " pod="openshift-nmstate/nmstate-handler-gnz9p" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.104487 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4plqx\" (UniqueName: \"kubernetes.io/projected/7e3e89d5-05d9-4675-9dad-83f59258f876-kube-api-access-4plqx\") pod \"nmstate-handler-gnz9p\" (UID: \"7e3e89d5-05d9-4675-9dad-83f59258f876\") " pod="openshift-nmstate/nmstate-handler-gnz9p" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.104607 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7wxv\" (UniqueName: \"kubernetes.io/projected/e047fa81-9835-4bca-95c4-c8d24f97857b-kube-api-access-n7wxv\") pod \"nmstate-webhook-6d689559c5-h7fmf\" (UID: \"e047fa81-9835-4bca-95c4-c8d24f97857b\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-h7fmf" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.104603 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/7e3e89d5-05d9-4675-9dad-83f59258f876-dbus-socket\") pod \"nmstate-handler-gnz9p\" (UID: \"7e3e89d5-05d9-4675-9dad-83f59258f876\") " pod="openshift-nmstate/nmstate-handler-gnz9p" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.122060 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4plqx\" (UniqueName: \"kubernetes.io/projected/7e3e89d5-05d9-4675-9dad-83f59258f876-kube-api-access-4plqx\") pod \"nmstate-handler-gnz9p\" (UID: \"7e3e89d5-05d9-4675-9dad-83f59258f876\") " pod="openshift-nmstate/nmstate-handler-gnz9p" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.205462 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7wxv\" (UniqueName: \"kubernetes.io/projected/e047fa81-9835-4bca-95c4-c8d24f97857b-kube-api-access-n7wxv\") pod \"nmstate-webhook-6d689559c5-h7fmf\" (UID: \"e047fa81-9835-4bca-95c4-c8d24f97857b\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-h7fmf" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.205538 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/14ccc5bb-b26a-42df-b0b9-f1ea9bcbff31-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-lw82x\" (UID: \"14ccc5bb-b26a-42df-b0b9-f1ea9bcbff31\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-lw82x" Sep 30 14:07:21 crc kubenswrapper[4840]: E0930 14:07:21.205679 4840 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.205700 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-957q6\" (UniqueName: \"kubernetes.io/projected/9ed06022-f9fe-4b61-bd42-ed308118a717-kube-api-access-957q6\") pod \"nmstate-metrics-58fcddf996-bw56r\" (UID: \"9ed06022-f9fe-4b61-bd42-ed308118a717\") " pod="openshift-nmstate/nmstate-metrics-58fcddf996-bw56r" Sep 30 14:07:21 crc kubenswrapper[4840]: E0930 14:07:21.205756 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/14ccc5bb-b26a-42df-b0b9-f1ea9bcbff31-plugin-serving-cert podName:14ccc5bb-b26a-42df-b0b9-f1ea9bcbff31 nodeName:}" failed. No retries permitted until 2025-09-30 14:07:21.705730817 +0000 UTC m=+670.334817240 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/14ccc5bb-b26a-42df-b0b9-f1ea9bcbff31-plugin-serving-cert") pod "nmstate-console-plugin-864bb6dfb5-lw82x" (UID: "14ccc5bb-b26a-42df-b0b9-f1ea9bcbff31") : secret "plugin-serving-cert" not found Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.205850 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/e047fa81-9835-4bca-95c4-c8d24f97857b-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-h7fmf\" (UID: \"e047fa81-9835-4bca-95c4-c8d24f97857b\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-h7fmf" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.205882 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/14ccc5bb-b26a-42df-b0b9-f1ea9bcbff31-nginx-conf\") pod \"nmstate-console-plugin-864bb6dfb5-lw82x\" (UID: \"14ccc5bb-b26a-42df-b0b9-f1ea9bcbff31\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-lw82x" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.205915 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9gzx\" (UniqueName: \"kubernetes.io/projected/14ccc5bb-b26a-42df-b0b9-f1ea9bcbff31-kube-api-access-v9gzx\") pod \"nmstate-console-plugin-864bb6dfb5-lw82x\" (UID: \"14ccc5bb-b26a-42df-b0b9-f1ea9bcbff31\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-lw82x" Sep 30 14:07:21 crc kubenswrapper[4840]: E0930 14:07:21.206491 4840 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Sep 30 14:07:21 crc kubenswrapper[4840]: E0930 14:07:21.206537 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e047fa81-9835-4bca-95c4-c8d24f97857b-tls-key-pair podName:e047fa81-9835-4bca-95c4-c8d24f97857b nodeName:}" failed. No retries permitted until 2025-09-30 14:07:21.706526177 +0000 UTC m=+670.335612600 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/e047fa81-9835-4bca-95c4-c8d24f97857b-tls-key-pair") pod "nmstate-webhook-6d689559c5-h7fmf" (UID: "e047fa81-9835-4bca-95c4-c8d24f97857b") : secret "openshift-nmstate-webhook" not found Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.207312 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/14ccc5bb-b26a-42df-b0b9-f1ea9bcbff31-nginx-conf\") pod \"nmstate-console-plugin-864bb6dfb5-lw82x\" (UID: \"14ccc5bb-b26a-42df-b0b9-f1ea9bcbff31\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-lw82x" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.227133 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9gzx\" (UniqueName: \"kubernetes.io/projected/14ccc5bb-b26a-42df-b0b9-f1ea9bcbff31-kube-api-access-v9gzx\") pod \"nmstate-console-plugin-864bb6dfb5-lw82x\" (UID: \"14ccc5bb-b26a-42df-b0b9-f1ea9bcbff31\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-lw82x" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.227422 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-957q6\" (UniqueName: \"kubernetes.io/projected/9ed06022-f9fe-4b61-bd42-ed308118a717-kube-api-access-957q6\") pod \"nmstate-metrics-58fcddf996-bw56r\" (UID: \"9ed06022-f9fe-4b61-bd42-ed308118a717\") " pod="openshift-nmstate/nmstate-metrics-58fcddf996-bw56r" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.230602 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7wxv\" (UniqueName: \"kubernetes.io/projected/e047fa81-9835-4bca-95c4-c8d24f97857b-kube-api-access-n7wxv\") pod \"nmstate-webhook-6d689559c5-h7fmf\" (UID: \"e047fa81-9835-4bca-95c4-c8d24f97857b\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-h7fmf" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.255503 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58fcddf996-bw56r" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.275695 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-76fd5f8594-tmhrt"] Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.276486 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-76fd5f8594-tmhrt" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.290322 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-gnz9p" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.308621 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/40167a8f-a497-44f0-bc03-4558335c2b8e-oauth-serving-cert\") pod \"console-76fd5f8594-tmhrt\" (UID: \"40167a8f-a497-44f0-bc03-4558335c2b8e\") " pod="openshift-console/console-76fd5f8594-tmhrt" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.308741 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/40167a8f-a497-44f0-bc03-4558335c2b8e-trusted-ca-bundle\") pod \"console-76fd5f8594-tmhrt\" (UID: \"40167a8f-a497-44f0-bc03-4558335c2b8e\") " pod="openshift-console/console-76fd5f8594-tmhrt" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.308766 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/40167a8f-a497-44f0-bc03-4558335c2b8e-service-ca\") pod \"console-76fd5f8594-tmhrt\" (UID: \"40167a8f-a497-44f0-bc03-4558335c2b8e\") " pod="openshift-console/console-76fd5f8594-tmhrt" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.308791 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tmc6\" (UniqueName: \"kubernetes.io/projected/40167a8f-a497-44f0-bc03-4558335c2b8e-kube-api-access-9tmc6\") pod \"console-76fd5f8594-tmhrt\" (UID: \"40167a8f-a497-44f0-bc03-4558335c2b8e\") " pod="openshift-console/console-76fd5f8594-tmhrt" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.308817 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/40167a8f-a497-44f0-bc03-4558335c2b8e-console-serving-cert\") pod \"console-76fd5f8594-tmhrt\" (UID: \"40167a8f-a497-44f0-bc03-4558335c2b8e\") " pod="openshift-console/console-76fd5f8594-tmhrt" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.308842 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/40167a8f-a497-44f0-bc03-4558335c2b8e-console-oauth-config\") pod \"console-76fd5f8594-tmhrt\" (UID: \"40167a8f-a497-44f0-bc03-4558335c2b8e\") " pod="openshift-console/console-76fd5f8594-tmhrt" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.308886 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/40167a8f-a497-44f0-bc03-4558335c2b8e-console-config\") pod \"console-76fd5f8594-tmhrt\" (UID: \"40167a8f-a497-44f0-bc03-4558335c2b8e\") " pod="openshift-console/console-76fd5f8594-tmhrt" Sep 30 14:07:21 crc kubenswrapper[4840]: W0930 14:07:21.318419 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e3e89d5_05d9_4675_9dad_83f59258f876.slice/crio-181e5cdbccb86ccf9f07b7752edd59e17d1816d08ea70643968c54337c8c5d79 WatchSource:0}: Error finding container 181e5cdbccb86ccf9f07b7752edd59e17d1816d08ea70643968c54337c8c5d79: Status 404 returned error can't find the container with id 181e5cdbccb86ccf9f07b7752edd59e17d1816d08ea70643968c54337c8c5d79 Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.330817 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-76fd5f8594-tmhrt"] Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.409860 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/40167a8f-a497-44f0-bc03-4558335c2b8e-trusted-ca-bundle\") pod \"console-76fd5f8594-tmhrt\" (UID: \"40167a8f-a497-44f0-bc03-4558335c2b8e\") " pod="openshift-console/console-76fd5f8594-tmhrt" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.409904 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/40167a8f-a497-44f0-bc03-4558335c2b8e-service-ca\") pod \"console-76fd5f8594-tmhrt\" (UID: \"40167a8f-a497-44f0-bc03-4558335c2b8e\") " pod="openshift-console/console-76fd5f8594-tmhrt" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.409923 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tmc6\" (UniqueName: \"kubernetes.io/projected/40167a8f-a497-44f0-bc03-4558335c2b8e-kube-api-access-9tmc6\") pod \"console-76fd5f8594-tmhrt\" (UID: \"40167a8f-a497-44f0-bc03-4558335c2b8e\") " pod="openshift-console/console-76fd5f8594-tmhrt" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.409948 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/40167a8f-a497-44f0-bc03-4558335c2b8e-console-serving-cert\") pod \"console-76fd5f8594-tmhrt\" (UID: \"40167a8f-a497-44f0-bc03-4558335c2b8e\") " pod="openshift-console/console-76fd5f8594-tmhrt" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.409963 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/40167a8f-a497-44f0-bc03-4558335c2b8e-console-oauth-config\") pod \"console-76fd5f8594-tmhrt\" (UID: \"40167a8f-a497-44f0-bc03-4558335c2b8e\") " pod="openshift-console/console-76fd5f8594-tmhrt" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.409989 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/40167a8f-a497-44f0-bc03-4558335c2b8e-console-config\") pod \"console-76fd5f8594-tmhrt\" (UID: \"40167a8f-a497-44f0-bc03-4558335c2b8e\") " pod="openshift-console/console-76fd5f8594-tmhrt" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.410019 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/40167a8f-a497-44f0-bc03-4558335c2b8e-oauth-serving-cert\") pod \"console-76fd5f8594-tmhrt\" (UID: \"40167a8f-a497-44f0-bc03-4558335c2b8e\") " pod="openshift-console/console-76fd5f8594-tmhrt" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.411142 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/40167a8f-a497-44f0-bc03-4558335c2b8e-oauth-serving-cert\") pod \"console-76fd5f8594-tmhrt\" (UID: \"40167a8f-a497-44f0-bc03-4558335c2b8e\") " pod="openshift-console/console-76fd5f8594-tmhrt" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.411339 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/40167a8f-a497-44f0-bc03-4558335c2b8e-trusted-ca-bundle\") pod \"console-76fd5f8594-tmhrt\" (UID: \"40167a8f-a497-44f0-bc03-4558335c2b8e\") " pod="openshift-console/console-76fd5f8594-tmhrt" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.412473 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/40167a8f-a497-44f0-bc03-4558335c2b8e-console-config\") pod \"console-76fd5f8594-tmhrt\" (UID: \"40167a8f-a497-44f0-bc03-4558335c2b8e\") " pod="openshift-console/console-76fd5f8594-tmhrt" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.413662 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/40167a8f-a497-44f0-bc03-4558335c2b8e-service-ca\") pod \"console-76fd5f8594-tmhrt\" (UID: \"40167a8f-a497-44f0-bc03-4558335c2b8e\") " pod="openshift-console/console-76fd5f8594-tmhrt" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.417288 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/40167a8f-a497-44f0-bc03-4558335c2b8e-console-oauth-config\") pod \"console-76fd5f8594-tmhrt\" (UID: \"40167a8f-a497-44f0-bc03-4558335c2b8e\") " pod="openshift-console/console-76fd5f8594-tmhrt" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.417910 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/40167a8f-a497-44f0-bc03-4558335c2b8e-console-serving-cert\") pod \"console-76fd5f8594-tmhrt\" (UID: \"40167a8f-a497-44f0-bc03-4558335c2b8e\") " pod="openshift-console/console-76fd5f8594-tmhrt" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.428435 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tmc6\" (UniqueName: \"kubernetes.io/projected/40167a8f-a497-44f0-bc03-4558335c2b8e-kube-api-access-9tmc6\") pod \"console-76fd5f8594-tmhrt\" (UID: \"40167a8f-a497-44f0-bc03-4558335c2b8e\") " pod="openshift-console/console-76fd5f8594-tmhrt" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.499937 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58fcddf996-bw56r"] Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.630994 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-76fd5f8594-tmhrt" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.713951 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/14ccc5bb-b26a-42df-b0b9-f1ea9bcbff31-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-lw82x\" (UID: \"14ccc5bb-b26a-42df-b0b9-f1ea9bcbff31\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-lw82x" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.714071 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/e047fa81-9835-4bca-95c4-c8d24f97857b-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-h7fmf\" (UID: \"e047fa81-9835-4bca-95c4-c8d24f97857b\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-h7fmf" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.718837 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/e047fa81-9835-4bca-95c4-c8d24f97857b-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-h7fmf\" (UID: \"e047fa81-9835-4bca-95c4-c8d24f97857b\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-h7fmf" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.718706 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/14ccc5bb-b26a-42df-b0b9-f1ea9bcbff31-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-lw82x\" (UID: \"14ccc5bb-b26a-42df-b0b9-f1ea9bcbff31\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-lw82x" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.812765 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-76fd5f8594-tmhrt"] Sep 30 14:07:21 crc kubenswrapper[4840]: W0930 14:07:21.819480 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod40167a8f_a497_44f0_bc03_4558335c2b8e.slice/crio-ba08cdbe1efd3901d5b314fcd06a7512eb15d3ffb419b8a60b9e8893daabd78b WatchSource:0}: Error finding container ba08cdbe1efd3901d5b314fcd06a7512eb15d3ffb419b8a60b9e8893daabd78b: Status 404 returned error can't find the container with id ba08cdbe1efd3901d5b314fcd06a7512eb15d3ffb419b8a60b9e8893daabd78b Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.870327 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6d689559c5-h7fmf" Sep 30 14:07:21 crc kubenswrapper[4840]: I0930 14:07:21.998077 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-lw82x" Sep 30 14:07:22 crc kubenswrapper[4840]: I0930 14:07:22.018537 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-76fd5f8594-tmhrt" event={"ID":"40167a8f-a497-44f0-bc03-4558335c2b8e","Type":"ContainerStarted","Data":"55f3007194281345f56ae39c5a96b45e97b74ff80070e714060c8d4959f5d639"} Sep 30 14:07:22 crc kubenswrapper[4840]: I0930 14:07:22.018824 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-76fd5f8594-tmhrt" event={"ID":"40167a8f-a497-44f0-bc03-4558335c2b8e","Type":"ContainerStarted","Data":"ba08cdbe1efd3901d5b314fcd06a7512eb15d3ffb419b8a60b9e8893daabd78b"} Sep 30 14:07:22 crc kubenswrapper[4840]: I0930 14:07:22.020405 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-gnz9p" event={"ID":"7e3e89d5-05d9-4675-9dad-83f59258f876","Type":"ContainerStarted","Data":"181e5cdbccb86ccf9f07b7752edd59e17d1816d08ea70643968c54337c8c5d79"} Sep 30 14:07:22 crc kubenswrapper[4840]: I0930 14:07:22.021788 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58fcddf996-bw56r" event={"ID":"9ed06022-f9fe-4b61-bd42-ed308118a717","Type":"ContainerStarted","Data":"4127b1d7cb8f869f82de549bf70c34bd15fa2fa33f0f71d7a99ff774da4d8313"} Sep 30 14:07:22 crc kubenswrapper[4840]: I0930 14:07:22.039620 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6d689559c5-h7fmf"] Sep 30 14:07:22 crc kubenswrapper[4840]: I0930 14:07:22.041535 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-76fd5f8594-tmhrt" podStartSLOduration=1.041515814 podStartE2EDuration="1.041515814s" podCreationTimestamp="2025-09-30 14:07:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:07:22.034225539 +0000 UTC m=+670.663311962" watchObservedRunningTime="2025-09-30 14:07:22.041515814 +0000 UTC m=+670.670602237" Sep 30 14:07:22 crc kubenswrapper[4840]: I0930 14:07:22.193038 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-864bb6dfb5-lw82x"] Sep 30 14:07:22 crc kubenswrapper[4840]: W0930 14:07:22.201533 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod14ccc5bb_b26a_42df_b0b9_f1ea9bcbff31.slice/crio-4349707f575fd8711e318907536e73455368e67c3982ad195816afb88c0eca03 WatchSource:0}: Error finding container 4349707f575fd8711e318907536e73455368e67c3982ad195816afb88c0eca03: Status 404 returned error can't find the container with id 4349707f575fd8711e318907536e73455368e67c3982ad195816afb88c0eca03 Sep 30 14:07:23 crc kubenswrapper[4840]: I0930 14:07:23.028282 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-lw82x" event={"ID":"14ccc5bb-b26a-42df-b0b9-f1ea9bcbff31","Type":"ContainerStarted","Data":"4349707f575fd8711e318907536e73455368e67c3982ad195816afb88c0eca03"} Sep 30 14:07:23 crc kubenswrapper[4840]: I0930 14:07:23.029605 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6d689559c5-h7fmf" event={"ID":"e047fa81-9835-4bca-95c4-c8d24f97857b","Type":"ContainerStarted","Data":"9f55a129d99e81e780d8dba9659df500b9c2917466524948722012df60ce5d41"} Sep 30 14:07:24 crc kubenswrapper[4840]: I0930 14:07:24.035468 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6d689559c5-h7fmf" event={"ID":"e047fa81-9835-4bca-95c4-c8d24f97857b","Type":"ContainerStarted","Data":"1e30121023085b51f760330487df6080109fdb7be5694dbd5458fbf92c64c5e7"} Sep 30 14:07:24 crc kubenswrapper[4840]: I0930 14:07:24.035886 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6d689559c5-h7fmf" Sep 30 14:07:24 crc kubenswrapper[4840]: I0930 14:07:24.037728 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58fcddf996-bw56r" event={"ID":"9ed06022-f9fe-4b61-bd42-ed308118a717","Type":"ContainerStarted","Data":"bd8247ea5eb50624dc35c23f83eea039c86e59f3a06e2a5971396c1d33ce98f6"} Sep 30 14:07:24 crc kubenswrapper[4840]: I0930 14:07:24.039183 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-gnz9p" event={"ID":"7e3e89d5-05d9-4675-9dad-83f59258f876","Type":"ContainerStarted","Data":"d91ce41b466a63e9a878b232a9b60853201f1123a7097996daa7284198603e02"} Sep 30 14:07:24 crc kubenswrapper[4840]: I0930 14:07:24.039323 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-gnz9p" Sep 30 14:07:24 crc kubenswrapper[4840]: I0930 14:07:24.055162 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6d689559c5-h7fmf" podStartSLOduration=2.380336773 podStartE2EDuration="4.05513857s" podCreationTimestamp="2025-09-30 14:07:20 +0000 UTC" firstStartedPulling="2025-09-30 14:07:22.051693922 +0000 UTC m=+670.680780345" lastFinishedPulling="2025-09-30 14:07:23.726495709 +0000 UTC m=+672.355582142" observedRunningTime="2025-09-30 14:07:24.051382984 +0000 UTC m=+672.680469407" watchObservedRunningTime="2025-09-30 14:07:24.05513857 +0000 UTC m=+672.684224993" Sep 30 14:07:24 crc kubenswrapper[4840]: I0930 14:07:24.071624 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-gnz9p" podStartSLOduration=1.6747367149999999 podStartE2EDuration="4.071604258s" podCreationTimestamp="2025-09-30 14:07:20 +0000 UTC" firstStartedPulling="2025-09-30 14:07:21.327650125 +0000 UTC m=+669.956736548" lastFinishedPulling="2025-09-30 14:07:23.724517648 +0000 UTC m=+672.353604091" observedRunningTime="2025-09-30 14:07:24.062824555 +0000 UTC m=+672.691910988" watchObservedRunningTime="2025-09-30 14:07:24.071604258 +0000 UTC m=+672.700690681" Sep 30 14:07:26 crc kubenswrapper[4840]: I0930 14:07:26.052591 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-lw82x" event={"ID":"14ccc5bb-b26a-42df-b0b9-f1ea9bcbff31","Type":"ContainerStarted","Data":"f1fc5e6270581683f66756717bf1c6ee135a0f6e595ce8bc0b3ee17f2c7be1e3"} Sep 30 14:07:26 crc kubenswrapper[4840]: I0930 14:07:26.054242 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58fcddf996-bw56r" event={"ID":"9ed06022-f9fe-4b61-bd42-ed308118a717","Type":"ContainerStarted","Data":"9a45f92864c16f766c901628ab12b224303dc5310b063895509cceddbbffa25e"} Sep 30 14:07:26 crc kubenswrapper[4840]: I0930 14:07:26.068808 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-lw82x" podStartSLOduration=2.3071575429999998 podStartE2EDuration="5.068790236s" podCreationTimestamp="2025-09-30 14:07:21 +0000 UTC" firstStartedPulling="2025-09-30 14:07:22.204230618 +0000 UTC m=+670.833317041" lastFinishedPulling="2025-09-30 14:07:24.965863311 +0000 UTC m=+673.594949734" observedRunningTime="2025-09-30 14:07:26.065246146 +0000 UTC m=+674.694332569" watchObservedRunningTime="2025-09-30 14:07:26.068790236 +0000 UTC m=+674.697876659" Sep 30 14:07:26 crc kubenswrapper[4840]: I0930 14:07:26.080434 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-58fcddf996-bw56r" podStartSLOduration=1.845895703 podStartE2EDuration="6.080420141s" podCreationTimestamp="2025-09-30 14:07:20 +0000 UTC" firstStartedPulling="2025-09-30 14:07:21.507499535 +0000 UTC m=+670.136585958" lastFinishedPulling="2025-09-30 14:07:25.742023983 +0000 UTC m=+674.371110396" observedRunningTime="2025-09-30 14:07:26.079112908 +0000 UTC m=+674.708199351" watchObservedRunningTime="2025-09-30 14:07:26.080420141 +0000 UTC m=+674.709506564" Sep 30 14:07:31 crc kubenswrapper[4840]: I0930 14:07:31.322605 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-gnz9p" Sep 30 14:07:31 crc kubenswrapper[4840]: I0930 14:07:31.631300 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-76fd5f8594-tmhrt" Sep 30 14:07:31 crc kubenswrapper[4840]: I0930 14:07:31.631350 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-76fd5f8594-tmhrt" Sep 30 14:07:31 crc kubenswrapper[4840]: I0930 14:07:31.637974 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-76fd5f8594-tmhrt" Sep 30 14:07:32 crc kubenswrapper[4840]: I0930 14:07:32.102631 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-76fd5f8594-tmhrt" Sep 30 14:07:32 crc kubenswrapper[4840]: I0930 14:07:32.180661 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-qjxnf"] Sep 30 14:07:41 crc kubenswrapper[4840]: I0930 14:07:41.878661 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6d689559c5-h7fmf" Sep 30 14:07:54 crc kubenswrapper[4840]: I0930 14:07:54.396133 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j"] Sep 30 14:07:54 crc kubenswrapper[4840]: I0930 14:07:54.397918 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j" Sep 30 14:07:54 crc kubenswrapper[4840]: I0930 14:07:54.399883 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Sep 30 14:07:54 crc kubenswrapper[4840]: I0930 14:07:54.409330 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j"] Sep 30 14:07:54 crc kubenswrapper[4840]: I0930 14:07:54.538059 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e8ac07f8-b0b5-4ff0-bae8-4d6594567c32-bundle\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j\" (UID: \"e8ac07f8-b0b5-4ff0-bae8-4d6594567c32\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j" Sep 30 14:07:54 crc kubenswrapper[4840]: I0930 14:07:54.538169 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e8ac07f8-b0b5-4ff0-bae8-4d6594567c32-util\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j\" (UID: \"e8ac07f8-b0b5-4ff0-bae8-4d6594567c32\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j" Sep 30 14:07:54 crc kubenswrapper[4840]: I0930 14:07:54.538227 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbhjd\" (UniqueName: \"kubernetes.io/projected/e8ac07f8-b0b5-4ff0-bae8-4d6594567c32-kube-api-access-pbhjd\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j\" (UID: \"e8ac07f8-b0b5-4ff0-bae8-4d6594567c32\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j" Sep 30 14:07:54 crc kubenswrapper[4840]: I0930 14:07:54.640292 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e8ac07f8-b0b5-4ff0-bae8-4d6594567c32-bundle\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j\" (UID: \"e8ac07f8-b0b5-4ff0-bae8-4d6594567c32\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j" Sep 30 14:07:54 crc kubenswrapper[4840]: I0930 14:07:54.640395 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e8ac07f8-b0b5-4ff0-bae8-4d6594567c32-util\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j\" (UID: \"e8ac07f8-b0b5-4ff0-bae8-4d6594567c32\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j" Sep 30 14:07:54 crc kubenswrapper[4840]: I0930 14:07:54.640436 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbhjd\" (UniqueName: \"kubernetes.io/projected/e8ac07f8-b0b5-4ff0-bae8-4d6594567c32-kube-api-access-pbhjd\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j\" (UID: \"e8ac07f8-b0b5-4ff0-bae8-4d6594567c32\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j" Sep 30 14:07:54 crc kubenswrapper[4840]: I0930 14:07:54.641321 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e8ac07f8-b0b5-4ff0-bae8-4d6594567c32-util\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j\" (UID: \"e8ac07f8-b0b5-4ff0-bae8-4d6594567c32\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j" Sep 30 14:07:54 crc kubenswrapper[4840]: I0930 14:07:54.641324 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e8ac07f8-b0b5-4ff0-bae8-4d6594567c32-bundle\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j\" (UID: \"e8ac07f8-b0b5-4ff0-bae8-4d6594567c32\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j" Sep 30 14:07:54 crc kubenswrapper[4840]: I0930 14:07:54.674239 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbhjd\" (UniqueName: \"kubernetes.io/projected/e8ac07f8-b0b5-4ff0-bae8-4d6594567c32-kube-api-access-pbhjd\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j\" (UID: \"e8ac07f8-b0b5-4ff0-bae8-4d6594567c32\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j" Sep 30 14:07:54 crc kubenswrapper[4840]: I0930 14:07:54.727427 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j" Sep 30 14:07:55 crc kubenswrapper[4840]: I0930 14:07:55.159165 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j"] Sep 30 14:07:55 crc kubenswrapper[4840]: I0930 14:07:55.234736 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j" event={"ID":"e8ac07f8-b0b5-4ff0-bae8-4d6594567c32","Type":"ContainerStarted","Data":"6dceb1ae88dd5b3b71e165a7f2de15e4f0bc9e0f4c7e05d9749b576c90b23b2c"} Sep 30 14:07:56 crc kubenswrapper[4840]: I0930 14:07:56.243630 4840 generic.go:334] "Generic (PLEG): container finished" podID="e8ac07f8-b0b5-4ff0-bae8-4d6594567c32" containerID="6e4445e2a6d70e737635b4974a72d5079e73c57d94e0d5c90cffa6e04015e6a7" exitCode=0 Sep 30 14:07:56 crc kubenswrapper[4840]: I0930 14:07:56.243674 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j" event={"ID":"e8ac07f8-b0b5-4ff0-bae8-4d6594567c32","Type":"ContainerDied","Data":"6e4445e2a6d70e737635b4974a72d5079e73c57d94e0d5c90cffa6e04015e6a7"} Sep 30 14:07:57 crc kubenswrapper[4840]: I0930 14:07:57.230052 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-qjxnf" podUID="9dd57b46-77fc-4b8a-9fc1-a50084165914" containerName="console" containerID="cri-o://175520c065fff8b959d46113bca4e5564c481dbced39acbba4f757dc0970dfff" gracePeriod=15 Sep 30 14:07:57 crc kubenswrapper[4840]: I0930 14:07:57.590022 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-qjxnf_9dd57b46-77fc-4b8a-9fc1-a50084165914/console/0.log" Sep 30 14:07:57 crc kubenswrapper[4840]: I0930 14:07:57.590306 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-qjxnf" Sep 30 14:07:57 crc kubenswrapper[4840]: I0930 14:07:57.682828 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9dd57b46-77fc-4b8a-9fc1-a50084165914-console-config\") pod \"9dd57b46-77fc-4b8a-9fc1-a50084165914\" (UID: \"9dd57b46-77fc-4b8a-9fc1-a50084165914\") " Sep 30 14:07:57 crc kubenswrapper[4840]: I0930 14:07:57.682895 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9dd57b46-77fc-4b8a-9fc1-a50084165914-console-oauth-config\") pod \"9dd57b46-77fc-4b8a-9fc1-a50084165914\" (UID: \"9dd57b46-77fc-4b8a-9fc1-a50084165914\") " Sep 30 14:07:57 crc kubenswrapper[4840]: I0930 14:07:57.682933 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9dd57b46-77fc-4b8a-9fc1-a50084165914-service-ca\") pod \"9dd57b46-77fc-4b8a-9fc1-a50084165914\" (UID: \"9dd57b46-77fc-4b8a-9fc1-a50084165914\") " Sep 30 14:07:57 crc kubenswrapper[4840]: I0930 14:07:57.682952 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9dd57b46-77fc-4b8a-9fc1-a50084165914-oauth-serving-cert\") pod \"9dd57b46-77fc-4b8a-9fc1-a50084165914\" (UID: \"9dd57b46-77fc-4b8a-9fc1-a50084165914\") " Sep 30 14:07:57 crc kubenswrapper[4840]: I0930 14:07:57.683027 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9dd57b46-77fc-4b8a-9fc1-a50084165914-console-serving-cert\") pod \"9dd57b46-77fc-4b8a-9fc1-a50084165914\" (UID: \"9dd57b46-77fc-4b8a-9fc1-a50084165914\") " Sep 30 14:07:57 crc kubenswrapper[4840]: I0930 14:07:57.683049 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpmvq\" (UniqueName: \"kubernetes.io/projected/9dd57b46-77fc-4b8a-9fc1-a50084165914-kube-api-access-cpmvq\") pod \"9dd57b46-77fc-4b8a-9fc1-a50084165914\" (UID: \"9dd57b46-77fc-4b8a-9fc1-a50084165914\") " Sep 30 14:07:57 crc kubenswrapper[4840]: I0930 14:07:57.683064 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9dd57b46-77fc-4b8a-9fc1-a50084165914-trusted-ca-bundle\") pod \"9dd57b46-77fc-4b8a-9fc1-a50084165914\" (UID: \"9dd57b46-77fc-4b8a-9fc1-a50084165914\") " Sep 30 14:07:57 crc kubenswrapper[4840]: I0930 14:07:57.683882 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9dd57b46-77fc-4b8a-9fc1-a50084165914-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "9dd57b46-77fc-4b8a-9fc1-a50084165914" (UID: "9dd57b46-77fc-4b8a-9fc1-a50084165914"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:07:57 crc kubenswrapper[4840]: I0930 14:07:57.683894 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9dd57b46-77fc-4b8a-9fc1-a50084165914-service-ca" (OuterVolumeSpecName: "service-ca") pod "9dd57b46-77fc-4b8a-9fc1-a50084165914" (UID: "9dd57b46-77fc-4b8a-9fc1-a50084165914"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:07:57 crc kubenswrapper[4840]: I0930 14:07:57.683933 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9dd57b46-77fc-4b8a-9fc1-a50084165914-console-config" (OuterVolumeSpecName: "console-config") pod "9dd57b46-77fc-4b8a-9fc1-a50084165914" (UID: "9dd57b46-77fc-4b8a-9fc1-a50084165914"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:07:57 crc kubenswrapper[4840]: I0930 14:07:57.684352 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9dd57b46-77fc-4b8a-9fc1-a50084165914-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "9dd57b46-77fc-4b8a-9fc1-a50084165914" (UID: "9dd57b46-77fc-4b8a-9fc1-a50084165914"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:07:57 crc kubenswrapper[4840]: I0930 14:07:57.688644 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9dd57b46-77fc-4b8a-9fc1-a50084165914-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "9dd57b46-77fc-4b8a-9fc1-a50084165914" (UID: "9dd57b46-77fc-4b8a-9fc1-a50084165914"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:07:57 crc kubenswrapper[4840]: I0930 14:07:57.689040 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9dd57b46-77fc-4b8a-9fc1-a50084165914-kube-api-access-cpmvq" (OuterVolumeSpecName: "kube-api-access-cpmvq") pod "9dd57b46-77fc-4b8a-9fc1-a50084165914" (UID: "9dd57b46-77fc-4b8a-9fc1-a50084165914"). InnerVolumeSpecName "kube-api-access-cpmvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:07:57 crc kubenswrapper[4840]: I0930 14:07:57.689467 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9dd57b46-77fc-4b8a-9fc1-a50084165914-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "9dd57b46-77fc-4b8a-9fc1-a50084165914" (UID: "9dd57b46-77fc-4b8a-9fc1-a50084165914"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:07:57 crc kubenswrapper[4840]: I0930 14:07:57.784265 4840 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9dd57b46-77fc-4b8a-9fc1-a50084165914-console-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:07:57 crc kubenswrapper[4840]: I0930 14:07:57.784295 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpmvq\" (UniqueName: \"kubernetes.io/projected/9dd57b46-77fc-4b8a-9fc1-a50084165914-kube-api-access-cpmvq\") on node \"crc\" DevicePath \"\"" Sep 30 14:07:57 crc kubenswrapper[4840]: I0930 14:07:57.784306 4840 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9dd57b46-77fc-4b8a-9fc1-a50084165914-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:07:57 crc kubenswrapper[4840]: I0930 14:07:57.784317 4840 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9dd57b46-77fc-4b8a-9fc1-a50084165914-console-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:07:57 crc kubenswrapper[4840]: I0930 14:07:57.784326 4840 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9dd57b46-77fc-4b8a-9fc1-a50084165914-console-oauth-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:07:57 crc kubenswrapper[4840]: I0930 14:07:57.784335 4840 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9dd57b46-77fc-4b8a-9fc1-a50084165914-service-ca\") on node \"crc\" DevicePath \"\"" Sep 30 14:07:57 crc kubenswrapper[4840]: I0930 14:07:57.784343 4840 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9dd57b46-77fc-4b8a-9fc1-a50084165914-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:07:58 crc kubenswrapper[4840]: I0930 14:07:58.258096 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-qjxnf_9dd57b46-77fc-4b8a-9fc1-a50084165914/console/0.log" Sep 30 14:07:58 crc kubenswrapper[4840]: I0930 14:07:58.258157 4840 generic.go:334] "Generic (PLEG): container finished" podID="9dd57b46-77fc-4b8a-9fc1-a50084165914" containerID="175520c065fff8b959d46113bca4e5564c481dbced39acbba4f757dc0970dfff" exitCode=2 Sep 30 14:07:58 crc kubenswrapper[4840]: I0930 14:07:58.258230 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-qjxnf" event={"ID":"9dd57b46-77fc-4b8a-9fc1-a50084165914","Type":"ContainerDied","Data":"175520c065fff8b959d46113bca4e5564c481dbced39acbba4f757dc0970dfff"} Sep 30 14:07:58 crc kubenswrapper[4840]: I0930 14:07:58.258245 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-qjxnf" Sep 30 14:07:58 crc kubenswrapper[4840]: I0930 14:07:58.258273 4840 scope.go:117] "RemoveContainer" containerID="175520c065fff8b959d46113bca4e5564c481dbced39acbba4f757dc0970dfff" Sep 30 14:07:58 crc kubenswrapper[4840]: I0930 14:07:58.258260 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-qjxnf" event={"ID":"9dd57b46-77fc-4b8a-9fc1-a50084165914","Type":"ContainerDied","Data":"d9412d6bc11d9a6ba329c315f49c51762fe98e4ea2c43e53b72fa624abc09c85"} Sep 30 14:07:58 crc kubenswrapper[4840]: I0930 14:07:58.260594 4840 generic.go:334] "Generic (PLEG): container finished" podID="e8ac07f8-b0b5-4ff0-bae8-4d6594567c32" containerID="8a2fe8ecbdb14dfc5d1155c629881dac00bfd039cbb4283e3744c624601b7421" exitCode=0 Sep 30 14:07:58 crc kubenswrapper[4840]: I0930 14:07:58.260637 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j" event={"ID":"e8ac07f8-b0b5-4ff0-bae8-4d6594567c32","Type":"ContainerDied","Data":"8a2fe8ecbdb14dfc5d1155c629881dac00bfd039cbb4283e3744c624601b7421"} Sep 30 14:07:58 crc kubenswrapper[4840]: I0930 14:07:58.286138 4840 scope.go:117] "RemoveContainer" containerID="175520c065fff8b959d46113bca4e5564c481dbced39acbba4f757dc0970dfff" Sep 30 14:07:58 crc kubenswrapper[4840]: E0930 14:07:58.286622 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"175520c065fff8b959d46113bca4e5564c481dbced39acbba4f757dc0970dfff\": container with ID starting with 175520c065fff8b959d46113bca4e5564c481dbced39acbba4f757dc0970dfff not found: ID does not exist" containerID="175520c065fff8b959d46113bca4e5564c481dbced39acbba4f757dc0970dfff" Sep 30 14:07:58 crc kubenswrapper[4840]: I0930 14:07:58.286674 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"175520c065fff8b959d46113bca4e5564c481dbced39acbba4f757dc0970dfff"} err="failed to get container status \"175520c065fff8b959d46113bca4e5564c481dbced39acbba4f757dc0970dfff\": rpc error: code = NotFound desc = could not find container \"175520c065fff8b959d46113bca4e5564c481dbced39acbba4f757dc0970dfff\": container with ID starting with 175520c065fff8b959d46113bca4e5564c481dbced39acbba4f757dc0970dfff not found: ID does not exist" Sep 30 14:07:58 crc kubenswrapper[4840]: I0930 14:07:58.307211 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-qjxnf"] Sep 30 14:07:58 crc kubenswrapper[4840]: I0930 14:07:58.311673 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-qjxnf"] Sep 30 14:07:59 crc kubenswrapper[4840]: I0930 14:07:59.270263 4840 generic.go:334] "Generic (PLEG): container finished" podID="e8ac07f8-b0b5-4ff0-bae8-4d6594567c32" containerID="d36628253c814b4394006e5a343b47f45feb53a5fca776dba029e482baf86685" exitCode=0 Sep 30 14:07:59 crc kubenswrapper[4840]: I0930 14:07:59.270337 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j" event={"ID":"e8ac07f8-b0b5-4ff0-bae8-4d6594567c32","Type":"ContainerDied","Data":"d36628253c814b4394006e5a343b47f45feb53a5fca776dba029e482baf86685"} Sep 30 14:08:00 crc kubenswrapper[4840]: I0930 14:08:00.135929 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9dd57b46-77fc-4b8a-9fc1-a50084165914" path="/var/lib/kubelet/pods/9dd57b46-77fc-4b8a-9fc1-a50084165914/volumes" Sep 30 14:08:00 crc kubenswrapper[4840]: I0930 14:08:00.555968 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j" Sep 30 14:08:00 crc kubenswrapper[4840]: I0930 14:08:00.720667 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pbhjd\" (UniqueName: \"kubernetes.io/projected/e8ac07f8-b0b5-4ff0-bae8-4d6594567c32-kube-api-access-pbhjd\") pod \"e8ac07f8-b0b5-4ff0-bae8-4d6594567c32\" (UID: \"e8ac07f8-b0b5-4ff0-bae8-4d6594567c32\") " Sep 30 14:08:00 crc kubenswrapper[4840]: I0930 14:08:00.720732 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e8ac07f8-b0b5-4ff0-bae8-4d6594567c32-util\") pod \"e8ac07f8-b0b5-4ff0-bae8-4d6594567c32\" (UID: \"e8ac07f8-b0b5-4ff0-bae8-4d6594567c32\") " Sep 30 14:08:00 crc kubenswrapper[4840]: I0930 14:08:00.720827 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e8ac07f8-b0b5-4ff0-bae8-4d6594567c32-bundle\") pod \"e8ac07f8-b0b5-4ff0-bae8-4d6594567c32\" (UID: \"e8ac07f8-b0b5-4ff0-bae8-4d6594567c32\") " Sep 30 14:08:00 crc kubenswrapper[4840]: I0930 14:08:00.722343 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8ac07f8-b0b5-4ff0-bae8-4d6594567c32-bundle" (OuterVolumeSpecName: "bundle") pod "e8ac07f8-b0b5-4ff0-bae8-4d6594567c32" (UID: "e8ac07f8-b0b5-4ff0-bae8-4d6594567c32"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:08:00 crc kubenswrapper[4840]: I0930 14:08:00.729837 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8ac07f8-b0b5-4ff0-bae8-4d6594567c32-kube-api-access-pbhjd" (OuterVolumeSpecName: "kube-api-access-pbhjd") pod "e8ac07f8-b0b5-4ff0-bae8-4d6594567c32" (UID: "e8ac07f8-b0b5-4ff0-bae8-4d6594567c32"). InnerVolumeSpecName "kube-api-access-pbhjd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:08:00 crc kubenswrapper[4840]: I0930 14:08:00.823112 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pbhjd\" (UniqueName: \"kubernetes.io/projected/e8ac07f8-b0b5-4ff0-bae8-4d6594567c32-kube-api-access-pbhjd\") on node \"crc\" DevicePath \"\"" Sep 30 14:08:00 crc kubenswrapper[4840]: I0930 14:08:00.823457 4840 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e8ac07f8-b0b5-4ff0-bae8-4d6594567c32-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:08:01 crc kubenswrapper[4840]: I0930 14:08:01.148840 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8ac07f8-b0b5-4ff0-bae8-4d6594567c32-util" (OuterVolumeSpecName: "util") pod "e8ac07f8-b0b5-4ff0-bae8-4d6594567c32" (UID: "e8ac07f8-b0b5-4ff0-bae8-4d6594567c32"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:08:01 crc kubenswrapper[4840]: I0930 14:08:01.230238 4840 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e8ac07f8-b0b5-4ff0-bae8-4d6594567c32-util\") on node \"crc\" DevicePath \"\"" Sep 30 14:08:01 crc kubenswrapper[4840]: I0930 14:08:01.287522 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j" event={"ID":"e8ac07f8-b0b5-4ff0-bae8-4d6594567c32","Type":"ContainerDied","Data":"6dceb1ae88dd5b3b71e165a7f2de15e4f0bc9e0f4c7e05d9749b576c90b23b2c"} Sep 30 14:08:01 crc kubenswrapper[4840]: I0930 14:08:01.287628 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6dceb1ae88dd5b3b71e165a7f2de15e4f0bc9e0f4c7e05d9749b576c90b23b2c" Sep 30 14:08:01 crc kubenswrapper[4840]: I0930 14:08:01.287710 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j" Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.273660 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-6c8fd557c5-29r6x"] Sep 30 14:08:09 crc kubenswrapper[4840]: E0930 14:08:09.274074 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8ac07f8-b0b5-4ff0-bae8-4d6594567c32" containerName="extract" Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.274086 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8ac07f8-b0b5-4ff0-bae8-4d6594567c32" containerName="extract" Sep 30 14:08:09 crc kubenswrapper[4840]: E0930 14:08:09.274097 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8ac07f8-b0b5-4ff0-bae8-4d6594567c32" containerName="util" Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.274103 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8ac07f8-b0b5-4ff0-bae8-4d6594567c32" containerName="util" Sep 30 14:08:09 crc kubenswrapper[4840]: E0930 14:08:09.274113 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8ac07f8-b0b5-4ff0-bae8-4d6594567c32" containerName="pull" Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.274119 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8ac07f8-b0b5-4ff0-bae8-4d6594567c32" containerName="pull" Sep 30 14:08:09 crc kubenswrapper[4840]: E0930 14:08:09.274128 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dd57b46-77fc-4b8a-9fc1-a50084165914" containerName="console" Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.274134 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dd57b46-77fc-4b8a-9fc1-a50084165914" containerName="console" Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.274223 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8ac07f8-b0b5-4ff0-bae8-4d6594567c32" containerName="extract" Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.274233 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dd57b46-77fc-4b8a-9fc1-a50084165914" containerName="console" Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.274595 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6c8fd557c5-29r6x" Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.275974 4840 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-8cdmt" Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.276010 4840 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.276332 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.276347 4840 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.276613 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.290852 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6c8fd557c5-29r6x"] Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.432309 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjtps\" (UniqueName: \"kubernetes.io/projected/ec2b2b7b-3b95-42c0-bfde-d8206541fe62-kube-api-access-jjtps\") pod \"metallb-operator-controller-manager-6c8fd557c5-29r6x\" (UID: \"ec2b2b7b-3b95-42c0-bfde-d8206541fe62\") " pod="metallb-system/metallb-operator-controller-manager-6c8fd557c5-29r6x" Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.432387 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ec2b2b7b-3b95-42c0-bfde-d8206541fe62-apiservice-cert\") pod \"metallb-operator-controller-manager-6c8fd557c5-29r6x\" (UID: \"ec2b2b7b-3b95-42c0-bfde-d8206541fe62\") " pod="metallb-system/metallb-operator-controller-manager-6c8fd557c5-29r6x" Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.432413 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ec2b2b7b-3b95-42c0-bfde-d8206541fe62-webhook-cert\") pod \"metallb-operator-controller-manager-6c8fd557c5-29r6x\" (UID: \"ec2b2b7b-3b95-42c0-bfde-d8206541fe62\") " pod="metallb-system/metallb-operator-controller-manager-6c8fd557c5-29r6x" Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.533110 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjtps\" (UniqueName: \"kubernetes.io/projected/ec2b2b7b-3b95-42c0-bfde-d8206541fe62-kube-api-access-jjtps\") pod \"metallb-operator-controller-manager-6c8fd557c5-29r6x\" (UID: \"ec2b2b7b-3b95-42c0-bfde-d8206541fe62\") " pod="metallb-system/metallb-operator-controller-manager-6c8fd557c5-29r6x" Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.533183 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ec2b2b7b-3b95-42c0-bfde-d8206541fe62-apiservice-cert\") pod \"metallb-operator-controller-manager-6c8fd557c5-29r6x\" (UID: \"ec2b2b7b-3b95-42c0-bfde-d8206541fe62\") " pod="metallb-system/metallb-operator-controller-manager-6c8fd557c5-29r6x" Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.533210 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ec2b2b7b-3b95-42c0-bfde-d8206541fe62-webhook-cert\") pod \"metallb-operator-controller-manager-6c8fd557c5-29r6x\" (UID: \"ec2b2b7b-3b95-42c0-bfde-d8206541fe62\") " pod="metallb-system/metallb-operator-controller-manager-6c8fd557c5-29r6x" Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.538300 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ec2b2b7b-3b95-42c0-bfde-d8206541fe62-apiservice-cert\") pod \"metallb-operator-controller-manager-6c8fd557c5-29r6x\" (UID: \"ec2b2b7b-3b95-42c0-bfde-d8206541fe62\") " pod="metallb-system/metallb-operator-controller-manager-6c8fd557c5-29r6x" Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.538801 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ec2b2b7b-3b95-42c0-bfde-d8206541fe62-webhook-cert\") pod \"metallb-operator-controller-manager-6c8fd557c5-29r6x\" (UID: \"ec2b2b7b-3b95-42c0-bfde-d8206541fe62\") " pod="metallb-system/metallb-operator-controller-manager-6c8fd557c5-29r6x" Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.560041 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjtps\" (UniqueName: \"kubernetes.io/projected/ec2b2b7b-3b95-42c0-bfde-d8206541fe62-kube-api-access-jjtps\") pod \"metallb-operator-controller-manager-6c8fd557c5-29r6x\" (UID: \"ec2b2b7b-3b95-42c0-bfde-d8206541fe62\") " pod="metallb-system/metallb-operator-controller-manager-6c8fd557c5-29r6x" Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.591538 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6c8fd557c5-29r6x" Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.732901 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-79c987c9f-v9dzf"] Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.738814 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-79c987c9f-v9dzf" Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.741689 4840 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-67tpx" Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.741931 4840 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.742066 4840 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.749879 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-79c987c9f-v9dzf"] Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.840170 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1e99fa8d-4919-483b-a71e-ee137198e55c-webhook-cert\") pod \"metallb-operator-webhook-server-79c987c9f-v9dzf\" (UID: \"1e99fa8d-4919-483b-a71e-ee137198e55c\") " pod="metallb-system/metallb-operator-webhook-server-79c987c9f-v9dzf" Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.840524 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1e99fa8d-4919-483b-a71e-ee137198e55c-apiservice-cert\") pod \"metallb-operator-webhook-server-79c987c9f-v9dzf\" (UID: \"1e99fa8d-4919-483b-a71e-ee137198e55c\") " pod="metallb-system/metallb-operator-webhook-server-79c987c9f-v9dzf" Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.840628 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x856w\" (UniqueName: \"kubernetes.io/projected/1e99fa8d-4919-483b-a71e-ee137198e55c-kube-api-access-x856w\") pod \"metallb-operator-webhook-server-79c987c9f-v9dzf\" (UID: \"1e99fa8d-4919-483b-a71e-ee137198e55c\") " pod="metallb-system/metallb-operator-webhook-server-79c987c9f-v9dzf" Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.842490 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6c8fd557c5-29r6x"] Sep 30 14:08:09 crc kubenswrapper[4840]: W0930 14:08:09.850859 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podec2b2b7b_3b95_42c0_bfde_d8206541fe62.slice/crio-df55d2cbc525183e23d4a59fcc2326bd36014fc30fc37ea63ca7e85fbbec180c WatchSource:0}: Error finding container df55d2cbc525183e23d4a59fcc2326bd36014fc30fc37ea63ca7e85fbbec180c: Status 404 returned error can't find the container with id df55d2cbc525183e23d4a59fcc2326bd36014fc30fc37ea63ca7e85fbbec180c Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.941671 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1e99fa8d-4919-483b-a71e-ee137198e55c-webhook-cert\") pod \"metallb-operator-webhook-server-79c987c9f-v9dzf\" (UID: \"1e99fa8d-4919-483b-a71e-ee137198e55c\") " pod="metallb-system/metallb-operator-webhook-server-79c987c9f-v9dzf" Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.941717 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1e99fa8d-4919-483b-a71e-ee137198e55c-apiservice-cert\") pod \"metallb-operator-webhook-server-79c987c9f-v9dzf\" (UID: \"1e99fa8d-4919-483b-a71e-ee137198e55c\") " pod="metallb-system/metallb-operator-webhook-server-79c987c9f-v9dzf" Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.941760 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x856w\" (UniqueName: \"kubernetes.io/projected/1e99fa8d-4919-483b-a71e-ee137198e55c-kube-api-access-x856w\") pod \"metallb-operator-webhook-server-79c987c9f-v9dzf\" (UID: \"1e99fa8d-4919-483b-a71e-ee137198e55c\") " pod="metallb-system/metallb-operator-webhook-server-79c987c9f-v9dzf" Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.947219 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1e99fa8d-4919-483b-a71e-ee137198e55c-apiservice-cert\") pod \"metallb-operator-webhook-server-79c987c9f-v9dzf\" (UID: \"1e99fa8d-4919-483b-a71e-ee137198e55c\") " pod="metallb-system/metallb-operator-webhook-server-79c987c9f-v9dzf" Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.947267 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1e99fa8d-4919-483b-a71e-ee137198e55c-webhook-cert\") pod \"metallb-operator-webhook-server-79c987c9f-v9dzf\" (UID: \"1e99fa8d-4919-483b-a71e-ee137198e55c\") " pod="metallb-system/metallb-operator-webhook-server-79c987c9f-v9dzf" Sep 30 14:08:09 crc kubenswrapper[4840]: I0930 14:08:09.964933 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x856w\" (UniqueName: \"kubernetes.io/projected/1e99fa8d-4919-483b-a71e-ee137198e55c-kube-api-access-x856w\") pod \"metallb-operator-webhook-server-79c987c9f-v9dzf\" (UID: \"1e99fa8d-4919-483b-a71e-ee137198e55c\") " pod="metallb-system/metallb-operator-webhook-server-79c987c9f-v9dzf" Sep 30 14:08:10 crc kubenswrapper[4840]: I0930 14:08:10.074376 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-79c987c9f-v9dzf" Sep 30 14:08:10 crc kubenswrapper[4840]: I0930 14:08:10.282413 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-79c987c9f-v9dzf"] Sep 30 14:08:10 crc kubenswrapper[4840]: W0930 14:08:10.290408 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1e99fa8d_4919_483b_a71e_ee137198e55c.slice/crio-4bfb66baf489da72f9c1e8d86b18b48c0f9c6a6ef45ec3224702aad6c29b6caa WatchSource:0}: Error finding container 4bfb66baf489da72f9c1e8d86b18b48c0f9c6a6ef45ec3224702aad6c29b6caa: Status 404 returned error can't find the container with id 4bfb66baf489da72f9c1e8d86b18b48c0f9c6a6ef45ec3224702aad6c29b6caa Sep 30 14:08:10 crc kubenswrapper[4840]: I0930 14:08:10.337262 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-79c987c9f-v9dzf" event={"ID":"1e99fa8d-4919-483b-a71e-ee137198e55c","Type":"ContainerStarted","Data":"4bfb66baf489da72f9c1e8d86b18b48c0f9c6a6ef45ec3224702aad6c29b6caa"} Sep 30 14:08:10 crc kubenswrapper[4840]: I0930 14:08:10.338335 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6c8fd557c5-29r6x" event={"ID":"ec2b2b7b-3b95-42c0-bfde-d8206541fe62","Type":"ContainerStarted","Data":"df55d2cbc525183e23d4a59fcc2326bd36014fc30fc37ea63ca7e85fbbec180c"} Sep 30 14:08:12 crc kubenswrapper[4840]: I0930 14:08:12.350244 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6c8fd557c5-29r6x" event={"ID":"ec2b2b7b-3b95-42c0-bfde-d8206541fe62","Type":"ContainerStarted","Data":"fa64b447c1411119da0d2757fa3a987b34664424a7d27746a1dd00e2a3bfdddd"} Sep 30 14:08:12 crc kubenswrapper[4840]: I0930 14:08:12.350459 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-6c8fd557c5-29r6x" Sep 30 14:08:12 crc kubenswrapper[4840]: I0930 14:08:12.382808 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-6c8fd557c5-29r6x" podStartSLOduration=1.208452774 podStartE2EDuration="3.382788159s" podCreationTimestamp="2025-09-30 14:08:09 +0000 UTC" firstStartedPulling="2025-09-30 14:08:09.852654351 +0000 UTC m=+718.481740774" lastFinishedPulling="2025-09-30 14:08:12.026989736 +0000 UTC m=+720.656076159" observedRunningTime="2025-09-30 14:08:12.378491339 +0000 UTC m=+721.007577772" watchObservedRunningTime="2025-09-30 14:08:12.382788159 +0000 UTC m=+721.011874592" Sep 30 14:08:15 crc kubenswrapper[4840]: I0930 14:08:15.369679 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-79c987c9f-v9dzf" event={"ID":"1e99fa8d-4919-483b-a71e-ee137198e55c","Type":"ContainerStarted","Data":"3021af141533710bdc3ca6c8ae05aed887d8d8be88b4a78fd030aad95ebe6c12"} Sep 30 14:08:15 crc kubenswrapper[4840]: I0930 14:08:15.370013 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-79c987c9f-v9dzf" Sep 30 14:08:15 crc kubenswrapper[4840]: I0930 14:08:15.398642 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-79c987c9f-v9dzf" podStartSLOduration=1.72980481 podStartE2EDuration="6.398627183s" podCreationTimestamp="2025-09-30 14:08:09 +0000 UTC" firstStartedPulling="2025-09-30 14:08:10.292972805 +0000 UTC m=+718.922059218" lastFinishedPulling="2025-09-30 14:08:14.961795168 +0000 UTC m=+723.590881591" observedRunningTime="2025-09-30 14:08:15.396010396 +0000 UTC m=+724.025096919" watchObservedRunningTime="2025-09-30 14:08:15.398627183 +0000 UTC m=+724.027713606" Sep 30 14:08:30 crc kubenswrapper[4840]: I0930 14:08:30.080103 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-79c987c9f-v9dzf" Sep 30 14:08:47 crc kubenswrapper[4840]: I0930 14:08:47.732479 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-wdjfz"] Sep 30 14:08:47 crc kubenswrapper[4840]: I0930 14:08:47.733188 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-wdjfz" podUID="661771aa-bd40-40fd-8a59-dc3b8156e5c5" containerName="controller-manager" containerID="cri-o://8425bcc99851578d42b5dc397401ae3091ff1c21fda2e7fe3746e0cd8dd51f8c" gracePeriod=30 Sep 30 14:08:47 crc kubenswrapper[4840]: I0930 14:08:47.823761 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftgw2"] Sep 30 14:08:47 crc kubenswrapper[4840]: I0930 14:08:47.823939 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftgw2" podUID="2842cd75-1ff5-4a3f-b7a0-8205388f4a17" containerName="route-controller-manager" containerID="cri-o://39c1272e3385ccb51e5260e86640e906d78ac3858150479e80f9fa91b40c984f" gracePeriod=30 Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.103884 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-wdjfz" Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.178020 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftgw2" Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.203797 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/661771aa-bd40-40fd-8a59-dc3b8156e5c5-proxy-ca-bundles\") pod \"661771aa-bd40-40fd-8a59-dc3b8156e5c5\" (UID: \"661771aa-bd40-40fd-8a59-dc3b8156e5c5\") " Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.203942 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/661771aa-bd40-40fd-8a59-dc3b8156e5c5-config\") pod \"661771aa-bd40-40fd-8a59-dc3b8156e5c5\" (UID: \"661771aa-bd40-40fd-8a59-dc3b8156e5c5\") " Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.203988 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/661771aa-bd40-40fd-8a59-dc3b8156e5c5-serving-cert\") pod \"661771aa-bd40-40fd-8a59-dc3b8156e5c5\" (UID: \"661771aa-bd40-40fd-8a59-dc3b8156e5c5\") " Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.204009 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/661771aa-bd40-40fd-8a59-dc3b8156e5c5-client-ca\") pod \"661771aa-bd40-40fd-8a59-dc3b8156e5c5\" (UID: \"661771aa-bd40-40fd-8a59-dc3b8156e5c5\") " Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.204051 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chhwt\" (UniqueName: \"kubernetes.io/projected/661771aa-bd40-40fd-8a59-dc3b8156e5c5-kube-api-access-chhwt\") pod \"661771aa-bd40-40fd-8a59-dc3b8156e5c5\" (UID: \"661771aa-bd40-40fd-8a59-dc3b8156e5c5\") " Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.204538 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/661771aa-bd40-40fd-8a59-dc3b8156e5c5-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "661771aa-bd40-40fd-8a59-dc3b8156e5c5" (UID: "661771aa-bd40-40fd-8a59-dc3b8156e5c5"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.205476 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/661771aa-bd40-40fd-8a59-dc3b8156e5c5-config" (OuterVolumeSpecName: "config") pod "661771aa-bd40-40fd-8a59-dc3b8156e5c5" (UID: "661771aa-bd40-40fd-8a59-dc3b8156e5c5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.205746 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/661771aa-bd40-40fd-8a59-dc3b8156e5c5-client-ca" (OuterVolumeSpecName: "client-ca") pod "661771aa-bd40-40fd-8a59-dc3b8156e5c5" (UID: "661771aa-bd40-40fd-8a59-dc3b8156e5c5"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.210627 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/661771aa-bd40-40fd-8a59-dc3b8156e5c5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "661771aa-bd40-40fd-8a59-dc3b8156e5c5" (UID: "661771aa-bd40-40fd-8a59-dc3b8156e5c5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.212051 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/661771aa-bd40-40fd-8a59-dc3b8156e5c5-kube-api-access-chhwt" (OuterVolumeSpecName: "kube-api-access-chhwt") pod "661771aa-bd40-40fd-8a59-dc3b8156e5c5" (UID: "661771aa-bd40-40fd-8a59-dc3b8156e5c5"). InnerVolumeSpecName "kube-api-access-chhwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.305723 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2842cd75-1ff5-4a3f-b7a0-8205388f4a17-client-ca\") pod \"2842cd75-1ff5-4a3f-b7a0-8205388f4a17\" (UID: \"2842cd75-1ff5-4a3f-b7a0-8205388f4a17\") " Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.305800 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2842cd75-1ff5-4a3f-b7a0-8205388f4a17-serving-cert\") pod \"2842cd75-1ff5-4a3f-b7a0-8205388f4a17\" (UID: \"2842cd75-1ff5-4a3f-b7a0-8205388f4a17\") " Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.305924 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2842cd75-1ff5-4a3f-b7a0-8205388f4a17-config\") pod \"2842cd75-1ff5-4a3f-b7a0-8205388f4a17\" (UID: \"2842cd75-1ff5-4a3f-b7a0-8205388f4a17\") " Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.305956 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvvvw\" (UniqueName: \"kubernetes.io/projected/2842cd75-1ff5-4a3f-b7a0-8205388f4a17-kube-api-access-vvvvw\") pod \"2842cd75-1ff5-4a3f-b7a0-8205388f4a17\" (UID: \"2842cd75-1ff5-4a3f-b7a0-8205388f4a17\") " Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.306191 4840 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/661771aa-bd40-40fd-8a59-dc3b8156e5c5-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.306207 4840 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/661771aa-bd40-40fd-8a59-dc3b8156e5c5-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.306219 4840 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/661771aa-bd40-40fd-8a59-dc3b8156e5c5-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.306230 4840 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/661771aa-bd40-40fd-8a59-dc3b8156e5c5-client-ca\") on node \"crc\" DevicePath \"\"" Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.306242 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chhwt\" (UniqueName: \"kubernetes.io/projected/661771aa-bd40-40fd-8a59-dc3b8156e5c5-kube-api-access-chhwt\") on node \"crc\" DevicePath \"\"" Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.306560 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2842cd75-1ff5-4a3f-b7a0-8205388f4a17-client-ca" (OuterVolumeSpecName: "client-ca") pod "2842cd75-1ff5-4a3f-b7a0-8205388f4a17" (UID: "2842cd75-1ff5-4a3f-b7a0-8205388f4a17"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.307252 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2842cd75-1ff5-4a3f-b7a0-8205388f4a17-config" (OuterVolumeSpecName: "config") pod "2842cd75-1ff5-4a3f-b7a0-8205388f4a17" (UID: "2842cd75-1ff5-4a3f-b7a0-8205388f4a17"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.309931 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2842cd75-1ff5-4a3f-b7a0-8205388f4a17-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "2842cd75-1ff5-4a3f-b7a0-8205388f4a17" (UID: "2842cd75-1ff5-4a3f-b7a0-8205388f4a17"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.310519 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2842cd75-1ff5-4a3f-b7a0-8205388f4a17-kube-api-access-vvvvw" (OuterVolumeSpecName: "kube-api-access-vvvvw") pod "2842cd75-1ff5-4a3f-b7a0-8205388f4a17" (UID: "2842cd75-1ff5-4a3f-b7a0-8205388f4a17"). InnerVolumeSpecName "kube-api-access-vvvvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.407631 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvvvw\" (UniqueName: \"kubernetes.io/projected/2842cd75-1ff5-4a3f-b7a0-8205388f4a17-kube-api-access-vvvvw\") on node \"crc\" DevicePath \"\"" Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.407664 4840 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2842cd75-1ff5-4a3f-b7a0-8205388f4a17-client-ca\") on node \"crc\" DevicePath \"\"" Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.407673 4840 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2842cd75-1ff5-4a3f-b7a0-8205388f4a17-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.407681 4840 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2842cd75-1ff5-4a3f-b7a0-8205388f4a17-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.537096 4840 generic.go:334] "Generic (PLEG): container finished" podID="2842cd75-1ff5-4a3f-b7a0-8205388f4a17" containerID="39c1272e3385ccb51e5260e86640e906d78ac3858150479e80f9fa91b40c984f" exitCode=0 Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.537191 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftgw2" event={"ID":"2842cd75-1ff5-4a3f-b7a0-8205388f4a17","Type":"ContainerDied","Data":"39c1272e3385ccb51e5260e86640e906d78ac3858150479e80f9fa91b40c984f"} Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.537216 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftgw2" Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.537245 4840 scope.go:117] "RemoveContainer" containerID="39c1272e3385ccb51e5260e86640e906d78ac3858150479e80f9fa91b40c984f" Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.537229 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftgw2" event={"ID":"2842cd75-1ff5-4a3f-b7a0-8205388f4a17","Type":"ContainerDied","Data":"e2521a8198db2bb35054a68c4f14f746a58a38ba30c028d7cd9c633400c0d9ef"} Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.539259 4840 generic.go:334] "Generic (PLEG): container finished" podID="661771aa-bd40-40fd-8a59-dc3b8156e5c5" containerID="8425bcc99851578d42b5dc397401ae3091ff1c21fda2e7fe3746e0cd8dd51f8c" exitCode=0 Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.539299 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-wdjfz" event={"ID":"661771aa-bd40-40fd-8a59-dc3b8156e5c5","Type":"ContainerDied","Data":"8425bcc99851578d42b5dc397401ae3091ff1c21fda2e7fe3746e0cd8dd51f8c"} Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.539317 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-wdjfz" event={"ID":"661771aa-bd40-40fd-8a59-dc3b8156e5c5","Type":"ContainerDied","Data":"6a56e0bfa14c27e4be3dc1dd85f6ea07a0b8d4189d81615588d76202ea3ef1a1"} Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.539325 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-wdjfz" Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.554173 4840 scope.go:117] "RemoveContainer" containerID="39c1272e3385ccb51e5260e86640e906d78ac3858150479e80f9fa91b40c984f" Sep 30 14:08:48 crc kubenswrapper[4840]: E0930 14:08:48.554864 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39c1272e3385ccb51e5260e86640e906d78ac3858150479e80f9fa91b40c984f\": container with ID starting with 39c1272e3385ccb51e5260e86640e906d78ac3858150479e80f9fa91b40c984f not found: ID does not exist" containerID="39c1272e3385ccb51e5260e86640e906d78ac3858150479e80f9fa91b40c984f" Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.554934 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39c1272e3385ccb51e5260e86640e906d78ac3858150479e80f9fa91b40c984f"} err="failed to get container status \"39c1272e3385ccb51e5260e86640e906d78ac3858150479e80f9fa91b40c984f\": rpc error: code = NotFound desc = could not find container \"39c1272e3385ccb51e5260e86640e906d78ac3858150479e80f9fa91b40c984f\": container with ID starting with 39c1272e3385ccb51e5260e86640e906d78ac3858150479e80f9fa91b40c984f not found: ID does not exist" Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.554978 4840 scope.go:117] "RemoveContainer" containerID="8425bcc99851578d42b5dc397401ae3091ff1c21fda2e7fe3746e0cd8dd51f8c" Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.568778 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-wdjfz"] Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.573558 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-wdjfz"] Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.576405 4840 scope.go:117] "RemoveContainer" containerID="8425bcc99851578d42b5dc397401ae3091ff1c21fda2e7fe3746e0cd8dd51f8c" Sep 30 14:08:48 crc kubenswrapper[4840]: E0930 14:08:48.577115 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8425bcc99851578d42b5dc397401ae3091ff1c21fda2e7fe3746e0cd8dd51f8c\": container with ID starting with 8425bcc99851578d42b5dc397401ae3091ff1c21fda2e7fe3746e0cd8dd51f8c not found: ID does not exist" containerID="8425bcc99851578d42b5dc397401ae3091ff1c21fda2e7fe3746e0cd8dd51f8c" Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.577148 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8425bcc99851578d42b5dc397401ae3091ff1c21fda2e7fe3746e0cd8dd51f8c"} err="failed to get container status \"8425bcc99851578d42b5dc397401ae3091ff1c21fda2e7fe3746e0cd8dd51f8c\": rpc error: code = NotFound desc = could not find container \"8425bcc99851578d42b5dc397401ae3091ff1c21fda2e7fe3746e0cd8dd51f8c\": container with ID starting with 8425bcc99851578d42b5dc397401ae3091ff1c21fda2e7fe3746e0cd8dd51f8c not found: ID does not exist" Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.580006 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftgw2"] Sep 30 14:08:48 crc kubenswrapper[4840]: I0930 14:08:48.584174 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ftgw2"] Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.133641 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-64557897d4-vnpxb"] Sep 30 14:08:49 crc kubenswrapper[4840]: E0930 14:08:49.134118 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="661771aa-bd40-40fd-8a59-dc3b8156e5c5" containerName="controller-manager" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.134130 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="661771aa-bd40-40fd-8a59-dc3b8156e5c5" containerName="controller-manager" Sep 30 14:08:49 crc kubenswrapper[4840]: E0930 14:08:49.134141 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2842cd75-1ff5-4a3f-b7a0-8205388f4a17" containerName="route-controller-manager" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.134148 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="2842cd75-1ff5-4a3f-b7a0-8205388f4a17" containerName="route-controller-manager" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.134248 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="2842cd75-1ff5-4a3f-b7a0-8205388f4a17" containerName="route-controller-manager" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.134260 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="661771aa-bd40-40fd-8a59-dc3b8156e5c5" containerName="controller-manager" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.134628 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-64557897d4-vnpxb" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.138077 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.138259 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.138077 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.138493 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.138156 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.138795 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.146197 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-64557897d4-vnpxb"] Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.217711 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/44735273-cc8d-4883-b3bf-3c815cfcc677-client-ca\") pod \"route-controller-manager-64557897d4-vnpxb\" (UID: \"44735273-cc8d-4883-b3bf-3c815cfcc677\") " pod="openshift-route-controller-manager/route-controller-manager-64557897d4-vnpxb" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.217783 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/44735273-cc8d-4883-b3bf-3c815cfcc677-serving-cert\") pod \"route-controller-manager-64557897d4-vnpxb\" (UID: \"44735273-cc8d-4883-b3bf-3c815cfcc677\") " pod="openshift-route-controller-manager/route-controller-manager-64557897d4-vnpxb" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.217917 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2ld7\" (UniqueName: \"kubernetes.io/projected/44735273-cc8d-4883-b3bf-3c815cfcc677-kube-api-access-v2ld7\") pod \"route-controller-manager-64557897d4-vnpxb\" (UID: \"44735273-cc8d-4883-b3bf-3c815cfcc677\") " pod="openshift-route-controller-manager/route-controller-manager-64557897d4-vnpxb" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.218017 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44735273-cc8d-4883-b3bf-3c815cfcc677-config\") pod \"route-controller-manager-64557897d4-vnpxb\" (UID: \"44735273-cc8d-4883-b3bf-3c815cfcc677\") " pod="openshift-route-controller-manager/route-controller-manager-64557897d4-vnpxb" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.318683 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/44735273-cc8d-4883-b3bf-3c815cfcc677-client-ca\") pod \"route-controller-manager-64557897d4-vnpxb\" (UID: \"44735273-cc8d-4883-b3bf-3c815cfcc677\") " pod="openshift-route-controller-manager/route-controller-manager-64557897d4-vnpxb" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.318741 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/44735273-cc8d-4883-b3bf-3c815cfcc677-serving-cert\") pod \"route-controller-manager-64557897d4-vnpxb\" (UID: \"44735273-cc8d-4883-b3bf-3c815cfcc677\") " pod="openshift-route-controller-manager/route-controller-manager-64557897d4-vnpxb" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.318798 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2ld7\" (UniqueName: \"kubernetes.io/projected/44735273-cc8d-4883-b3bf-3c815cfcc677-kube-api-access-v2ld7\") pod \"route-controller-manager-64557897d4-vnpxb\" (UID: \"44735273-cc8d-4883-b3bf-3c815cfcc677\") " pod="openshift-route-controller-manager/route-controller-manager-64557897d4-vnpxb" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.318833 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44735273-cc8d-4883-b3bf-3c815cfcc677-config\") pod \"route-controller-manager-64557897d4-vnpxb\" (UID: \"44735273-cc8d-4883-b3bf-3c815cfcc677\") " pod="openshift-route-controller-manager/route-controller-manager-64557897d4-vnpxb" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.319877 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44735273-cc8d-4883-b3bf-3c815cfcc677-config\") pod \"route-controller-manager-64557897d4-vnpxb\" (UID: \"44735273-cc8d-4883-b3bf-3c815cfcc677\") " pod="openshift-route-controller-manager/route-controller-manager-64557897d4-vnpxb" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.320166 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/44735273-cc8d-4883-b3bf-3c815cfcc677-client-ca\") pod \"route-controller-manager-64557897d4-vnpxb\" (UID: \"44735273-cc8d-4883-b3bf-3c815cfcc677\") " pod="openshift-route-controller-manager/route-controller-manager-64557897d4-vnpxb" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.324587 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/44735273-cc8d-4883-b3bf-3c815cfcc677-serving-cert\") pod \"route-controller-manager-64557897d4-vnpxb\" (UID: \"44735273-cc8d-4883-b3bf-3c815cfcc677\") " pod="openshift-route-controller-manager/route-controller-manager-64557897d4-vnpxb" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.335973 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2ld7\" (UniqueName: \"kubernetes.io/projected/44735273-cc8d-4883-b3bf-3c815cfcc677-kube-api-access-v2ld7\") pod \"route-controller-manager-64557897d4-vnpxb\" (UID: \"44735273-cc8d-4883-b3bf-3c815cfcc677\") " pod="openshift-route-controller-manager/route-controller-manager-64557897d4-vnpxb" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.446768 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-67f7cc6ddb-nznkr"] Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.447816 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-67f7cc6ddb-nznkr" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.450342 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.450446 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.450674 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.450757 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.451368 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-64557897d4-vnpxb" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.451496 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.453168 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.464208 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.466989 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-67f7cc6ddb-nznkr"] Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.521311 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/17bc05cf-564a-4401-981c-f050fa85f57a-client-ca\") pod \"controller-manager-67f7cc6ddb-nznkr\" (UID: \"17bc05cf-564a-4401-981c-f050fa85f57a\") " pod="openshift-controller-manager/controller-manager-67f7cc6ddb-nznkr" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.521588 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54j8n\" (UniqueName: \"kubernetes.io/projected/17bc05cf-564a-4401-981c-f050fa85f57a-kube-api-access-54j8n\") pod \"controller-manager-67f7cc6ddb-nznkr\" (UID: \"17bc05cf-564a-4401-981c-f050fa85f57a\") " pod="openshift-controller-manager/controller-manager-67f7cc6ddb-nznkr" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.521683 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17bc05cf-564a-4401-981c-f050fa85f57a-config\") pod \"controller-manager-67f7cc6ddb-nznkr\" (UID: \"17bc05cf-564a-4401-981c-f050fa85f57a\") " pod="openshift-controller-manager/controller-manager-67f7cc6ddb-nznkr" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.521754 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17bc05cf-564a-4401-981c-f050fa85f57a-serving-cert\") pod \"controller-manager-67f7cc6ddb-nznkr\" (UID: \"17bc05cf-564a-4401-981c-f050fa85f57a\") " pod="openshift-controller-manager/controller-manager-67f7cc6ddb-nznkr" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.521864 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/17bc05cf-564a-4401-981c-f050fa85f57a-proxy-ca-bundles\") pod \"controller-manager-67f7cc6ddb-nznkr\" (UID: \"17bc05cf-564a-4401-981c-f050fa85f57a\") " pod="openshift-controller-manager/controller-manager-67f7cc6ddb-nznkr" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.594212 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-6c8fd557c5-29r6x" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.623372 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17bc05cf-564a-4401-981c-f050fa85f57a-config\") pod \"controller-manager-67f7cc6ddb-nznkr\" (UID: \"17bc05cf-564a-4401-981c-f050fa85f57a\") " pod="openshift-controller-manager/controller-manager-67f7cc6ddb-nznkr" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.623425 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17bc05cf-564a-4401-981c-f050fa85f57a-serving-cert\") pod \"controller-manager-67f7cc6ddb-nznkr\" (UID: \"17bc05cf-564a-4401-981c-f050fa85f57a\") " pod="openshift-controller-manager/controller-manager-67f7cc6ddb-nznkr" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.623470 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/17bc05cf-564a-4401-981c-f050fa85f57a-proxy-ca-bundles\") pod \"controller-manager-67f7cc6ddb-nznkr\" (UID: \"17bc05cf-564a-4401-981c-f050fa85f57a\") " pod="openshift-controller-manager/controller-manager-67f7cc6ddb-nznkr" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.623522 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/17bc05cf-564a-4401-981c-f050fa85f57a-client-ca\") pod \"controller-manager-67f7cc6ddb-nznkr\" (UID: \"17bc05cf-564a-4401-981c-f050fa85f57a\") " pod="openshift-controller-manager/controller-manager-67f7cc6ddb-nznkr" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.623603 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54j8n\" (UniqueName: \"kubernetes.io/projected/17bc05cf-564a-4401-981c-f050fa85f57a-kube-api-access-54j8n\") pod \"controller-manager-67f7cc6ddb-nznkr\" (UID: \"17bc05cf-564a-4401-981c-f050fa85f57a\") " pod="openshift-controller-manager/controller-manager-67f7cc6ddb-nznkr" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.625328 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/17bc05cf-564a-4401-981c-f050fa85f57a-client-ca\") pod \"controller-manager-67f7cc6ddb-nznkr\" (UID: \"17bc05cf-564a-4401-981c-f050fa85f57a\") " pod="openshift-controller-manager/controller-manager-67f7cc6ddb-nznkr" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.626204 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/17bc05cf-564a-4401-981c-f050fa85f57a-proxy-ca-bundles\") pod \"controller-manager-67f7cc6ddb-nznkr\" (UID: \"17bc05cf-564a-4401-981c-f050fa85f57a\") " pod="openshift-controller-manager/controller-manager-67f7cc6ddb-nznkr" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.628535 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17bc05cf-564a-4401-981c-f050fa85f57a-config\") pod \"controller-manager-67f7cc6ddb-nznkr\" (UID: \"17bc05cf-564a-4401-981c-f050fa85f57a\") " pod="openshift-controller-manager/controller-manager-67f7cc6ddb-nznkr" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.633442 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17bc05cf-564a-4401-981c-f050fa85f57a-serving-cert\") pod \"controller-manager-67f7cc6ddb-nznkr\" (UID: \"17bc05cf-564a-4401-981c-f050fa85f57a\") " pod="openshift-controller-manager/controller-manager-67f7cc6ddb-nznkr" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.643574 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54j8n\" (UniqueName: \"kubernetes.io/projected/17bc05cf-564a-4401-981c-f050fa85f57a-kube-api-access-54j8n\") pod \"controller-manager-67f7cc6ddb-nznkr\" (UID: \"17bc05cf-564a-4401-981c-f050fa85f57a\") " pod="openshift-controller-manager/controller-manager-67f7cc6ddb-nznkr" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.687902 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-64557897d4-vnpxb"] Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.765130 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-67f7cc6ddb-nznkr" Sep 30 14:08:49 crc kubenswrapper[4840]: I0930 14:08:49.980413 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-67f7cc6ddb-nznkr"] Sep 30 14:08:49 crc kubenswrapper[4840]: W0930 14:08:49.989101 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17bc05cf_564a_4401_981c_f050fa85f57a.slice/crio-e03160f63eadaa257c98b3e3d2dc6deca7e1cd76ed2db1e7a46b1a46e7ff2aed WatchSource:0}: Error finding container e03160f63eadaa257c98b3e3d2dc6deca7e1cd76ed2db1e7a46b1a46e7ff2aed: Status 404 returned error can't find the container with id e03160f63eadaa257c98b3e3d2dc6deca7e1cd76ed2db1e7a46b1a46e7ff2aed Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.125482 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2842cd75-1ff5-4a3f-b7a0-8205388f4a17" path="/var/lib/kubelet/pods/2842cd75-1ff5-4a3f-b7a0-8205388f4a17/volumes" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.126157 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="661771aa-bd40-40fd-8a59-dc3b8156e5c5" path="/var/lib/kubelet/pods/661771aa-bd40-40fd-8a59-dc3b8156e5c5/volumes" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.320841 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-sqx2j"] Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.330186 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-8qmmm"] Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.330948 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-8qmmm" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.331478 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-sqx2j" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.332942 4840 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-kwvln" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.333291 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.333461 4840 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.335798 4840 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.343846 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-8qmmm"] Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.430510 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-hnv4n"] Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.433565 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-hnv4n" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.435732 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/624a00a9-8fda-4e40-9df0-e8452ed89d28-frr-sockets\") pod \"frr-k8s-sqx2j\" (UID: \"624a00a9-8fda-4e40-9df0-e8452ed89d28\") " pod="metallb-system/frr-k8s-sqx2j" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.435757 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/624a00a9-8fda-4e40-9df0-e8452ed89d28-metrics-certs\") pod \"frr-k8s-sqx2j\" (UID: \"624a00a9-8fda-4e40-9df0-e8452ed89d28\") " pod="metallb-system/frr-k8s-sqx2j" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.435784 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/624a00a9-8fda-4e40-9df0-e8452ed89d28-frr-startup\") pod \"frr-k8s-sqx2j\" (UID: \"624a00a9-8fda-4e40-9df0-e8452ed89d28\") " pod="metallb-system/frr-k8s-sqx2j" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.435812 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/624a00a9-8fda-4e40-9df0-e8452ed89d28-metrics\") pod \"frr-k8s-sqx2j\" (UID: \"624a00a9-8fda-4e40-9df0-e8452ed89d28\") " pod="metallb-system/frr-k8s-sqx2j" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.435832 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/624a00a9-8fda-4e40-9df0-e8452ed89d28-frr-conf\") pod \"frr-k8s-sqx2j\" (UID: \"624a00a9-8fda-4e40-9df0-e8452ed89d28\") " pod="metallb-system/frr-k8s-sqx2j" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.435883 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7348a607-eb78-4fbf-a620-9be8180b8503-cert\") pod \"frr-k8s-webhook-server-5478bdb765-8qmmm\" (UID: \"7348a607-eb78-4fbf-a620-9be8180b8503\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-8qmmm" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.435919 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/624a00a9-8fda-4e40-9df0-e8452ed89d28-reloader\") pod \"frr-k8s-sqx2j\" (UID: \"624a00a9-8fda-4e40-9df0-e8452ed89d28\") " pod="metallb-system/frr-k8s-sqx2j" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.435938 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79drx\" (UniqueName: \"kubernetes.io/projected/624a00a9-8fda-4e40-9df0-e8452ed89d28-kube-api-access-79drx\") pod \"frr-k8s-sqx2j\" (UID: \"624a00a9-8fda-4e40-9df0-e8452ed89d28\") " pod="metallb-system/frr-k8s-sqx2j" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.435960 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llsq7\" (UniqueName: \"kubernetes.io/projected/7348a607-eb78-4fbf-a620-9be8180b8503-kube-api-access-llsq7\") pod \"frr-k8s-webhook-server-5478bdb765-8qmmm\" (UID: \"7348a607-eb78-4fbf-a620-9be8180b8503\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-8qmmm" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.448835 4840 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-kjx5p" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.449065 4840 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.449567 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.453414 4840 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.459805 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-5d688f5ffc-nv2dz"] Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.460692 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5d688f5ffc-nv2dz" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.464566 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5d688f5ffc-nv2dz"] Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.465967 4840 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.538183 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/624a00a9-8fda-4e40-9df0-e8452ed89d28-metrics\") pod \"frr-k8s-sqx2j\" (UID: \"624a00a9-8fda-4e40-9df0-e8452ed89d28\") " pod="metallb-system/frr-k8s-sqx2j" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.538235 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/624a00a9-8fda-4e40-9df0-e8452ed89d28-frr-conf\") pod \"frr-k8s-sqx2j\" (UID: \"624a00a9-8fda-4e40-9df0-e8452ed89d28\") " pod="metallb-system/frr-k8s-sqx2j" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.538265 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e2682a09-9f63-42a1-882a-368be12b8be0-cert\") pod \"controller-5d688f5ffc-nv2dz\" (UID: \"e2682a09-9f63-42a1-882a-368be12b8be0\") " pod="metallb-system/controller-5d688f5ffc-nv2dz" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.538288 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/a15e1237-7226-4180-8c27-fb06dffbd2d7-metallb-excludel2\") pod \"speaker-hnv4n\" (UID: \"a15e1237-7226-4180-8c27-fb06dffbd2d7\") " pod="metallb-system/speaker-hnv4n" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.538328 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a15e1237-7226-4180-8c27-fb06dffbd2d7-memberlist\") pod \"speaker-hnv4n\" (UID: \"a15e1237-7226-4180-8c27-fb06dffbd2d7\") " pod="metallb-system/speaker-hnv4n" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.538351 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7348a607-eb78-4fbf-a620-9be8180b8503-cert\") pod \"frr-k8s-webhook-server-5478bdb765-8qmmm\" (UID: \"7348a607-eb78-4fbf-a620-9be8180b8503\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-8qmmm" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.538383 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/624a00a9-8fda-4e40-9df0-e8452ed89d28-reloader\") pod \"frr-k8s-sqx2j\" (UID: \"624a00a9-8fda-4e40-9df0-e8452ed89d28\") " pod="metallb-system/frr-k8s-sqx2j" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.538400 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79drx\" (UniqueName: \"kubernetes.io/projected/624a00a9-8fda-4e40-9df0-e8452ed89d28-kube-api-access-79drx\") pod \"frr-k8s-sqx2j\" (UID: \"624a00a9-8fda-4e40-9df0-e8452ed89d28\") " pod="metallb-system/frr-k8s-sqx2j" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.538418 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e2682a09-9f63-42a1-882a-368be12b8be0-metrics-certs\") pod \"controller-5d688f5ffc-nv2dz\" (UID: \"e2682a09-9f63-42a1-882a-368be12b8be0\") " pod="metallb-system/controller-5d688f5ffc-nv2dz" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.538436 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llsq7\" (UniqueName: \"kubernetes.io/projected/7348a607-eb78-4fbf-a620-9be8180b8503-kube-api-access-llsq7\") pod \"frr-k8s-webhook-server-5478bdb765-8qmmm\" (UID: \"7348a607-eb78-4fbf-a620-9be8180b8503\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-8qmmm" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.538470 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/624a00a9-8fda-4e40-9df0-e8452ed89d28-frr-sockets\") pod \"frr-k8s-sqx2j\" (UID: \"624a00a9-8fda-4e40-9df0-e8452ed89d28\") " pod="metallb-system/frr-k8s-sqx2j" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.538485 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/624a00a9-8fda-4e40-9df0-e8452ed89d28-metrics-certs\") pod \"frr-k8s-sqx2j\" (UID: \"624a00a9-8fda-4e40-9df0-e8452ed89d28\") " pod="metallb-system/frr-k8s-sqx2j" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.538505 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a15e1237-7226-4180-8c27-fb06dffbd2d7-metrics-certs\") pod \"speaker-hnv4n\" (UID: \"a15e1237-7226-4180-8c27-fb06dffbd2d7\") " pod="metallb-system/speaker-hnv4n" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.538524 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vgs8\" (UniqueName: \"kubernetes.io/projected/e2682a09-9f63-42a1-882a-368be12b8be0-kube-api-access-7vgs8\") pod \"controller-5d688f5ffc-nv2dz\" (UID: \"e2682a09-9f63-42a1-882a-368be12b8be0\") " pod="metallb-system/controller-5d688f5ffc-nv2dz" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.538540 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bj4mc\" (UniqueName: \"kubernetes.io/projected/a15e1237-7226-4180-8c27-fb06dffbd2d7-kube-api-access-bj4mc\") pod \"speaker-hnv4n\" (UID: \"a15e1237-7226-4180-8c27-fb06dffbd2d7\") " pod="metallb-system/speaker-hnv4n" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.538573 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/624a00a9-8fda-4e40-9df0-e8452ed89d28-frr-startup\") pod \"frr-k8s-sqx2j\" (UID: \"624a00a9-8fda-4e40-9df0-e8452ed89d28\") " pod="metallb-system/frr-k8s-sqx2j" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.539577 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/624a00a9-8fda-4e40-9df0-e8452ed89d28-frr-startup\") pod \"frr-k8s-sqx2j\" (UID: \"624a00a9-8fda-4e40-9df0-e8452ed89d28\") " pod="metallb-system/frr-k8s-sqx2j" Sep 30 14:08:50 crc kubenswrapper[4840]: E0930 14:08:50.539902 4840 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.539959 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/624a00a9-8fda-4e40-9df0-e8452ed89d28-metrics\") pod \"frr-k8s-sqx2j\" (UID: \"624a00a9-8fda-4e40-9df0-e8452ed89d28\") " pod="metallb-system/frr-k8s-sqx2j" Sep 30 14:08:50 crc kubenswrapper[4840]: E0930 14:08:50.539971 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7348a607-eb78-4fbf-a620-9be8180b8503-cert podName:7348a607-eb78-4fbf-a620-9be8180b8503 nodeName:}" failed. No retries permitted until 2025-09-30 14:08:51.039952555 +0000 UTC m=+759.669038988 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/7348a607-eb78-4fbf-a620-9be8180b8503-cert") pod "frr-k8s-webhook-server-5478bdb765-8qmmm" (UID: "7348a607-eb78-4fbf-a620-9be8180b8503") : secret "frr-k8s-webhook-server-cert" not found Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.540147 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/624a00a9-8fda-4e40-9df0-e8452ed89d28-frr-sockets\") pod \"frr-k8s-sqx2j\" (UID: \"624a00a9-8fda-4e40-9df0-e8452ed89d28\") " pod="metallb-system/frr-k8s-sqx2j" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.540201 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/624a00a9-8fda-4e40-9df0-e8452ed89d28-frr-conf\") pod \"frr-k8s-sqx2j\" (UID: \"624a00a9-8fda-4e40-9df0-e8452ed89d28\") " pod="metallb-system/frr-k8s-sqx2j" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.540413 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/624a00a9-8fda-4e40-9df0-e8452ed89d28-reloader\") pod \"frr-k8s-sqx2j\" (UID: \"624a00a9-8fda-4e40-9df0-e8452ed89d28\") " pod="metallb-system/frr-k8s-sqx2j" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.556454 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/624a00a9-8fda-4e40-9df0-e8452ed89d28-metrics-certs\") pod \"frr-k8s-sqx2j\" (UID: \"624a00a9-8fda-4e40-9df0-e8452ed89d28\") " pod="metallb-system/frr-k8s-sqx2j" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.572387 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-67f7cc6ddb-nznkr" event={"ID":"17bc05cf-564a-4401-981c-f050fa85f57a","Type":"ContainerStarted","Data":"3295baf63fbd0d908dc2473de0569007e066d2e46686b6a9be1fe12f26a13414"} Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.572443 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-67f7cc6ddb-nznkr" event={"ID":"17bc05cf-564a-4401-981c-f050fa85f57a","Type":"ContainerStarted","Data":"e03160f63eadaa257c98b3e3d2dc6deca7e1cd76ed2db1e7a46b1a46e7ff2aed"} Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.573427 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-67f7cc6ddb-nznkr" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.579724 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-64557897d4-vnpxb" event={"ID":"44735273-cc8d-4883-b3bf-3c815cfcc677","Type":"ContainerStarted","Data":"cd6b5cc4f06160aa53ac2058e394c863f5c29a1c0da0507835fffe6ee752a113"} Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.579769 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-64557897d4-vnpxb" event={"ID":"44735273-cc8d-4883-b3bf-3c815cfcc677","Type":"ContainerStarted","Data":"63c9d0269955b3ae680bdd78b54de7f3329fc411bd487825f45e4abc9e3353cd"} Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.584897 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-64557897d4-vnpxb" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.598567 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llsq7\" (UniqueName: \"kubernetes.io/projected/7348a607-eb78-4fbf-a620-9be8180b8503-kube-api-access-llsq7\") pod \"frr-k8s-webhook-server-5478bdb765-8qmmm\" (UID: \"7348a607-eb78-4fbf-a620-9be8180b8503\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-8qmmm" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.602640 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-67f7cc6ddb-nznkr" podStartSLOduration=3.602620257 podStartE2EDuration="3.602620257s" podCreationTimestamp="2025-09-30 14:08:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:08:50.599120048 +0000 UTC m=+759.228206471" watchObservedRunningTime="2025-09-30 14:08:50.602620257 +0000 UTC m=+759.231706680" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.604191 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-64557897d4-vnpxb" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.622896 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-67f7cc6ddb-nznkr" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.623218 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79drx\" (UniqueName: \"kubernetes.io/projected/624a00a9-8fda-4e40-9df0-e8452ed89d28-kube-api-access-79drx\") pod \"frr-k8s-sqx2j\" (UID: \"624a00a9-8fda-4e40-9df0-e8452ed89d28\") " pod="metallb-system/frr-k8s-sqx2j" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.639806 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a15e1237-7226-4180-8c27-fb06dffbd2d7-memberlist\") pod \"speaker-hnv4n\" (UID: \"a15e1237-7226-4180-8c27-fb06dffbd2d7\") " pod="metallb-system/speaker-hnv4n" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.639927 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e2682a09-9f63-42a1-882a-368be12b8be0-metrics-certs\") pod \"controller-5d688f5ffc-nv2dz\" (UID: \"e2682a09-9f63-42a1-882a-368be12b8be0\") " pod="metallb-system/controller-5d688f5ffc-nv2dz" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.639983 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a15e1237-7226-4180-8c27-fb06dffbd2d7-metrics-certs\") pod \"speaker-hnv4n\" (UID: \"a15e1237-7226-4180-8c27-fb06dffbd2d7\") " pod="metallb-system/speaker-hnv4n" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.640021 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vgs8\" (UniqueName: \"kubernetes.io/projected/e2682a09-9f63-42a1-882a-368be12b8be0-kube-api-access-7vgs8\") pod \"controller-5d688f5ffc-nv2dz\" (UID: \"e2682a09-9f63-42a1-882a-368be12b8be0\") " pod="metallb-system/controller-5d688f5ffc-nv2dz" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.640045 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bj4mc\" (UniqueName: \"kubernetes.io/projected/a15e1237-7226-4180-8c27-fb06dffbd2d7-kube-api-access-bj4mc\") pod \"speaker-hnv4n\" (UID: \"a15e1237-7226-4180-8c27-fb06dffbd2d7\") " pod="metallb-system/speaker-hnv4n" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.640078 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/a15e1237-7226-4180-8c27-fb06dffbd2d7-metallb-excludel2\") pod \"speaker-hnv4n\" (UID: \"a15e1237-7226-4180-8c27-fb06dffbd2d7\") " pod="metallb-system/speaker-hnv4n" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.640099 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e2682a09-9f63-42a1-882a-368be12b8be0-cert\") pod \"controller-5d688f5ffc-nv2dz\" (UID: \"e2682a09-9f63-42a1-882a-368be12b8be0\") " pod="metallb-system/controller-5d688f5ffc-nv2dz" Sep 30 14:08:50 crc kubenswrapper[4840]: E0930 14:08:50.640265 4840 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Sep 30 14:08:50 crc kubenswrapper[4840]: E0930 14:08:50.640323 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a15e1237-7226-4180-8c27-fb06dffbd2d7-memberlist podName:a15e1237-7226-4180-8c27-fb06dffbd2d7 nodeName:}" failed. No retries permitted until 2025-09-30 14:08:51.14030692 +0000 UTC m=+759.769393343 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/a15e1237-7226-4180-8c27-fb06dffbd2d7-memberlist") pod "speaker-hnv4n" (UID: "a15e1237-7226-4180-8c27-fb06dffbd2d7") : secret "metallb-memberlist" not found Sep 30 14:08:50 crc kubenswrapper[4840]: E0930 14:08:50.642026 4840 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Sep 30 14:08:50 crc kubenswrapper[4840]: E0930 14:08:50.642072 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e2682a09-9f63-42a1-882a-368be12b8be0-metrics-certs podName:e2682a09-9f63-42a1-882a-368be12b8be0 nodeName:}" failed. No retries permitted until 2025-09-30 14:08:51.142057465 +0000 UTC m=+759.771143888 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e2682a09-9f63-42a1-882a-368be12b8be0-metrics-certs") pod "controller-5d688f5ffc-nv2dz" (UID: "e2682a09-9f63-42a1-882a-368be12b8be0") : secret "controller-certs-secret" not found Sep 30 14:08:50 crc kubenswrapper[4840]: E0930 14:08:50.642298 4840 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Sep 30 14:08:50 crc kubenswrapper[4840]: E0930 14:08:50.642331 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a15e1237-7226-4180-8c27-fb06dffbd2d7-metrics-certs podName:a15e1237-7226-4180-8c27-fb06dffbd2d7 nodeName:}" failed. No retries permitted until 2025-09-30 14:08:51.142322112 +0000 UTC m=+759.771408535 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a15e1237-7226-4180-8c27-fb06dffbd2d7-metrics-certs") pod "speaker-hnv4n" (UID: "a15e1237-7226-4180-8c27-fb06dffbd2d7") : secret "speaker-certs-secret" not found Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.643049 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/a15e1237-7226-4180-8c27-fb06dffbd2d7-metallb-excludel2\") pod \"speaker-hnv4n\" (UID: \"a15e1237-7226-4180-8c27-fb06dffbd2d7\") " pod="metallb-system/speaker-hnv4n" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.650836 4840 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.655428 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e2682a09-9f63-42a1-882a-368be12b8be0-cert\") pod \"controller-5d688f5ffc-nv2dz\" (UID: \"e2682a09-9f63-42a1-882a-368be12b8be0\") " pod="metallb-system/controller-5d688f5ffc-nv2dz" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.655528 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-64557897d4-vnpxb" podStartSLOduration=1.655510049 podStartE2EDuration="1.655510049s" podCreationTimestamp="2025-09-30 14:08:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:08:50.651159738 +0000 UTC m=+759.280246161" watchObservedRunningTime="2025-09-30 14:08:50.655510049 +0000 UTC m=+759.284596472" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.665605 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-sqx2j" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.673219 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vgs8\" (UniqueName: \"kubernetes.io/projected/e2682a09-9f63-42a1-882a-368be12b8be0-kube-api-access-7vgs8\") pod \"controller-5d688f5ffc-nv2dz\" (UID: \"e2682a09-9f63-42a1-882a-368be12b8be0\") " pod="metallb-system/controller-5d688f5ffc-nv2dz" Sep 30 14:08:50 crc kubenswrapper[4840]: I0930 14:08:50.682105 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bj4mc\" (UniqueName: \"kubernetes.io/projected/a15e1237-7226-4180-8c27-fb06dffbd2d7-kube-api-access-bj4mc\") pod \"speaker-hnv4n\" (UID: \"a15e1237-7226-4180-8c27-fb06dffbd2d7\") " pod="metallb-system/speaker-hnv4n" Sep 30 14:08:51 crc kubenswrapper[4840]: I0930 14:08:51.045800 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7348a607-eb78-4fbf-a620-9be8180b8503-cert\") pod \"frr-k8s-webhook-server-5478bdb765-8qmmm\" (UID: \"7348a607-eb78-4fbf-a620-9be8180b8503\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-8qmmm" Sep 30 14:08:51 crc kubenswrapper[4840]: I0930 14:08:51.053407 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7348a607-eb78-4fbf-a620-9be8180b8503-cert\") pod \"frr-k8s-webhook-server-5478bdb765-8qmmm\" (UID: \"7348a607-eb78-4fbf-a620-9be8180b8503\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-8qmmm" Sep 30 14:08:51 crc kubenswrapper[4840]: I0930 14:08:51.146864 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a15e1237-7226-4180-8c27-fb06dffbd2d7-memberlist\") pod \"speaker-hnv4n\" (UID: \"a15e1237-7226-4180-8c27-fb06dffbd2d7\") " pod="metallb-system/speaker-hnv4n" Sep 30 14:08:51 crc kubenswrapper[4840]: I0930 14:08:51.146930 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e2682a09-9f63-42a1-882a-368be12b8be0-metrics-certs\") pod \"controller-5d688f5ffc-nv2dz\" (UID: \"e2682a09-9f63-42a1-882a-368be12b8be0\") " pod="metallb-system/controller-5d688f5ffc-nv2dz" Sep 30 14:08:51 crc kubenswrapper[4840]: I0930 14:08:51.146972 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a15e1237-7226-4180-8c27-fb06dffbd2d7-metrics-certs\") pod \"speaker-hnv4n\" (UID: \"a15e1237-7226-4180-8c27-fb06dffbd2d7\") " pod="metallb-system/speaker-hnv4n" Sep 30 14:08:51 crc kubenswrapper[4840]: E0930 14:08:51.147116 4840 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Sep 30 14:08:51 crc kubenswrapper[4840]: E0930 14:08:51.147204 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a15e1237-7226-4180-8c27-fb06dffbd2d7-memberlist podName:a15e1237-7226-4180-8c27-fb06dffbd2d7 nodeName:}" failed. No retries permitted until 2025-09-30 14:08:52.147179106 +0000 UTC m=+760.776265529 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/a15e1237-7226-4180-8c27-fb06dffbd2d7-memberlist") pod "speaker-hnv4n" (UID: "a15e1237-7226-4180-8c27-fb06dffbd2d7") : secret "metallb-memberlist" not found Sep 30 14:08:51 crc kubenswrapper[4840]: I0930 14:08:51.150229 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a15e1237-7226-4180-8c27-fb06dffbd2d7-metrics-certs\") pod \"speaker-hnv4n\" (UID: \"a15e1237-7226-4180-8c27-fb06dffbd2d7\") " pod="metallb-system/speaker-hnv4n" Sep 30 14:08:51 crc kubenswrapper[4840]: I0930 14:08:51.150713 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e2682a09-9f63-42a1-882a-368be12b8be0-metrics-certs\") pod \"controller-5d688f5ffc-nv2dz\" (UID: \"e2682a09-9f63-42a1-882a-368be12b8be0\") " pod="metallb-system/controller-5d688f5ffc-nv2dz" Sep 30 14:08:51 crc kubenswrapper[4840]: I0930 14:08:51.251847 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-8qmmm" Sep 30 14:08:51 crc kubenswrapper[4840]: I0930 14:08:51.403266 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5d688f5ffc-nv2dz" Sep 30 14:08:51 crc kubenswrapper[4840]: I0930 14:08:51.459521 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-8qmmm"] Sep 30 14:08:51 crc kubenswrapper[4840]: I0930 14:08:51.586348 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-8qmmm" event={"ID":"7348a607-eb78-4fbf-a620-9be8180b8503","Type":"ContainerStarted","Data":"9e57d11e2f7059525a7f509e2df014e68eda5921b2ee428487d87989e02d8d8f"} Sep 30 14:08:51 crc kubenswrapper[4840]: I0930 14:08:51.587192 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sqx2j" event={"ID":"624a00a9-8fda-4e40-9df0-e8452ed89d28","Type":"ContainerStarted","Data":"3b09e653dd03b6b3b27622c87d82b8c2725366c3a775897db3249be5792a09a7"} Sep 30 14:08:51 crc kubenswrapper[4840]: I0930 14:08:51.839738 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5d688f5ffc-nv2dz"] Sep 30 14:08:51 crc kubenswrapper[4840]: W0930 14:08:51.844786 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode2682a09_9f63_42a1_882a_368be12b8be0.slice/crio-ed7c63c671d34a614ba79ff893a94b682cd011de65f8f65e2d585924045b1f4c WatchSource:0}: Error finding container ed7c63c671d34a614ba79ff893a94b682cd011de65f8f65e2d585924045b1f4c: Status 404 returned error can't find the container with id ed7c63c671d34a614ba79ff893a94b682cd011de65f8f65e2d585924045b1f4c Sep 30 14:08:52 crc kubenswrapper[4840]: I0930 14:08:52.166326 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a15e1237-7226-4180-8c27-fb06dffbd2d7-memberlist\") pod \"speaker-hnv4n\" (UID: \"a15e1237-7226-4180-8c27-fb06dffbd2d7\") " pod="metallb-system/speaker-hnv4n" Sep 30 14:08:52 crc kubenswrapper[4840]: I0930 14:08:52.171990 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a15e1237-7226-4180-8c27-fb06dffbd2d7-memberlist\") pod \"speaker-hnv4n\" (UID: \"a15e1237-7226-4180-8c27-fb06dffbd2d7\") " pod="metallb-system/speaker-hnv4n" Sep 30 14:08:52 crc kubenswrapper[4840]: I0930 14:08:52.283886 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-hnv4n" Sep 30 14:08:52 crc kubenswrapper[4840]: W0930 14:08:52.303691 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda15e1237_7226_4180_8c27_fb06dffbd2d7.slice/crio-2463f00116de13925ef3b27b53713ad846d3ed26a50ffb99ad0ac38d8cd30af6 WatchSource:0}: Error finding container 2463f00116de13925ef3b27b53713ad846d3ed26a50ffb99ad0ac38d8cd30af6: Status 404 returned error can't find the container with id 2463f00116de13925ef3b27b53713ad846d3ed26a50ffb99ad0ac38d8cd30af6 Sep 30 14:08:52 crc kubenswrapper[4840]: I0930 14:08:52.595821 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5d688f5ffc-nv2dz" event={"ID":"e2682a09-9f63-42a1-882a-368be12b8be0","Type":"ContainerStarted","Data":"2cb82d5a6b802a5e5000c51a9c0eb59052cbe355b3d9172702fea97e46ebf1af"} Sep 30 14:08:52 crc kubenswrapper[4840]: I0930 14:08:52.596109 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5d688f5ffc-nv2dz" event={"ID":"e2682a09-9f63-42a1-882a-368be12b8be0","Type":"ContainerStarted","Data":"772759598c2ab83cae9ab10dc3ac4ecbec134876464698e55c89490df355517c"} Sep 30 14:08:52 crc kubenswrapper[4840]: I0930 14:08:52.596121 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5d688f5ffc-nv2dz" event={"ID":"e2682a09-9f63-42a1-882a-368be12b8be0","Type":"ContainerStarted","Data":"ed7c63c671d34a614ba79ff893a94b682cd011de65f8f65e2d585924045b1f4c"} Sep 30 14:08:52 crc kubenswrapper[4840]: I0930 14:08:52.596338 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-5d688f5ffc-nv2dz" Sep 30 14:08:52 crc kubenswrapper[4840]: I0930 14:08:52.605709 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-hnv4n" event={"ID":"a15e1237-7226-4180-8c27-fb06dffbd2d7","Type":"ContainerStarted","Data":"5c2f8b1bd5265f132c509ade214fd27f2e6ab928297be07d1c3b762a7ea05014"} Sep 30 14:08:52 crc kubenswrapper[4840]: I0930 14:08:52.605753 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-hnv4n" event={"ID":"a15e1237-7226-4180-8c27-fb06dffbd2d7","Type":"ContainerStarted","Data":"2463f00116de13925ef3b27b53713ad846d3ed26a50ffb99ad0ac38d8cd30af6"} Sep 30 14:08:52 crc kubenswrapper[4840]: I0930 14:08:52.609394 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-5d688f5ffc-nv2dz" podStartSLOduration=2.609379348 podStartE2EDuration="2.609379348s" podCreationTimestamp="2025-09-30 14:08:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:08:52.608593028 +0000 UTC m=+761.237679451" watchObservedRunningTime="2025-09-30 14:08:52.609379348 +0000 UTC m=+761.238465771" Sep 30 14:08:53 crc kubenswrapper[4840]: I0930 14:08:53.615235 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-hnv4n" event={"ID":"a15e1237-7226-4180-8c27-fb06dffbd2d7","Type":"ContainerStarted","Data":"cd4e70a435ad627717cdea30901c64b7c3bbb14a007d7d5c76b12fadaa8d8f40"} Sep 30 14:08:53 crc kubenswrapper[4840]: I0930 14:08:53.639544 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-hnv4n" podStartSLOduration=3.639527008 podStartE2EDuration="3.639527008s" podCreationTimestamp="2025-09-30 14:08:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:08:53.63917673 +0000 UTC m=+762.268263153" watchObservedRunningTime="2025-09-30 14:08:53.639527008 +0000 UTC m=+762.268613431" Sep 30 14:08:53 crc kubenswrapper[4840]: I0930 14:08:53.997056 4840 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Sep 30 14:08:54 crc kubenswrapper[4840]: I0930 14:08:54.622262 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-hnv4n" Sep 30 14:08:56 crc kubenswrapper[4840]: I0930 14:08:56.557849 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jhpfb"] Sep 30 14:08:56 crc kubenswrapper[4840]: I0930 14:08:56.559414 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jhpfb" Sep 30 14:08:56 crc kubenswrapper[4840]: I0930 14:08:56.571246 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jhpfb"] Sep 30 14:08:56 crc kubenswrapper[4840]: I0930 14:08:56.631898 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6480ac95-e05f-40d5-b6e2-934aa91f0eb8-utilities\") pod \"certified-operators-jhpfb\" (UID: \"6480ac95-e05f-40d5-b6e2-934aa91f0eb8\") " pod="openshift-marketplace/certified-operators-jhpfb" Sep 30 14:08:56 crc kubenswrapper[4840]: I0930 14:08:56.631964 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6mtw\" (UniqueName: \"kubernetes.io/projected/6480ac95-e05f-40d5-b6e2-934aa91f0eb8-kube-api-access-r6mtw\") pod \"certified-operators-jhpfb\" (UID: \"6480ac95-e05f-40d5-b6e2-934aa91f0eb8\") " pod="openshift-marketplace/certified-operators-jhpfb" Sep 30 14:08:56 crc kubenswrapper[4840]: I0930 14:08:56.631988 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6480ac95-e05f-40d5-b6e2-934aa91f0eb8-catalog-content\") pod \"certified-operators-jhpfb\" (UID: \"6480ac95-e05f-40d5-b6e2-934aa91f0eb8\") " pod="openshift-marketplace/certified-operators-jhpfb" Sep 30 14:08:56 crc kubenswrapper[4840]: I0930 14:08:56.733898 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6480ac95-e05f-40d5-b6e2-934aa91f0eb8-utilities\") pod \"certified-operators-jhpfb\" (UID: \"6480ac95-e05f-40d5-b6e2-934aa91f0eb8\") " pod="openshift-marketplace/certified-operators-jhpfb" Sep 30 14:08:56 crc kubenswrapper[4840]: I0930 14:08:56.733960 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6mtw\" (UniqueName: \"kubernetes.io/projected/6480ac95-e05f-40d5-b6e2-934aa91f0eb8-kube-api-access-r6mtw\") pod \"certified-operators-jhpfb\" (UID: \"6480ac95-e05f-40d5-b6e2-934aa91f0eb8\") " pod="openshift-marketplace/certified-operators-jhpfb" Sep 30 14:08:56 crc kubenswrapper[4840]: I0930 14:08:56.733980 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6480ac95-e05f-40d5-b6e2-934aa91f0eb8-catalog-content\") pod \"certified-operators-jhpfb\" (UID: \"6480ac95-e05f-40d5-b6e2-934aa91f0eb8\") " pod="openshift-marketplace/certified-operators-jhpfb" Sep 30 14:08:56 crc kubenswrapper[4840]: I0930 14:08:56.734487 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6480ac95-e05f-40d5-b6e2-934aa91f0eb8-utilities\") pod \"certified-operators-jhpfb\" (UID: \"6480ac95-e05f-40d5-b6e2-934aa91f0eb8\") " pod="openshift-marketplace/certified-operators-jhpfb" Sep 30 14:08:56 crc kubenswrapper[4840]: I0930 14:08:56.734504 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6480ac95-e05f-40d5-b6e2-934aa91f0eb8-catalog-content\") pod \"certified-operators-jhpfb\" (UID: \"6480ac95-e05f-40d5-b6e2-934aa91f0eb8\") " pod="openshift-marketplace/certified-operators-jhpfb" Sep 30 14:08:56 crc kubenswrapper[4840]: I0930 14:08:56.755349 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6mtw\" (UniqueName: \"kubernetes.io/projected/6480ac95-e05f-40d5-b6e2-934aa91f0eb8-kube-api-access-r6mtw\") pod \"certified-operators-jhpfb\" (UID: \"6480ac95-e05f-40d5-b6e2-934aa91f0eb8\") " pod="openshift-marketplace/certified-operators-jhpfb" Sep 30 14:08:56 crc kubenswrapper[4840]: I0930 14:08:56.883002 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jhpfb" Sep 30 14:08:58 crc kubenswrapper[4840]: I0930 14:08:58.848569 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jhpfb"] Sep 30 14:08:59 crc kubenswrapper[4840]: I0930 14:08:59.655417 4840 generic.go:334] "Generic (PLEG): container finished" podID="6480ac95-e05f-40d5-b6e2-934aa91f0eb8" containerID="392f8ef1ade10ace1c9b635fdb9932d87d576e38dc065e74c87893971f7efbe1" exitCode=0 Sep 30 14:08:59 crc kubenswrapper[4840]: I0930 14:08:59.655513 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jhpfb" event={"ID":"6480ac95-e05f-40d5-b6e2-934aa91f0eb8","Type":"ContainerDied","Data":"392f8ef1ade10ace1c9b635fdb9932d87d576e38dc065e74c87893971f7efbe1"} Sep 30 14:08:59 crc kubenswrapper[4840]: I0930 14:08:59.655789 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jhpfb" event={"ID":"6480ac95-e05f-40d5-b6e2-934aa91f0eb8","Type":"ContainerStarted","Data":"9d183a2457b3eabb33ee32715f1170f6e558f939df42a40b2c8f8415075c17f3"} Sep 30 14:08:59 crc kubenswrapper[4840]: I0930 14:08:59.657754 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-8qmmm" event={"ID":"7348a607-eb78-4fbf-a620-9be8180b8503","Type":"ContainerStarted","Data":"1252317295d8f6d986f84d37ef6e7ddddc1d7ba6b369a3716df31690dcd131f6"} Sep 30 14:08:59 crc kubenswrapper[4840]: I0930 14:08:59.657857 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-8qmmm" Sep 30 14:08:59 crc kubenswrapper[4840]: I0930 14:08:59.660390 4840 generic.go:334] "Generic (PLEG): container finished" podID="624a00a9-8fda-4e40-9df0-e8452ed89d28" containerID="88d4cd37deb09bfde951bdf429f702351e943de6aaaaccc55abf5775b28d7795" exitCode=0 Sep 30 14:08:59 crc kubenswrapper[4840]: I0930 14:08:59.660441 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sqx2j" event={"ID":"624a00a9-8fda-4e40-9df0-e8452ed89d28","Type":"ContainerDied","Data":"88d4cd37deb09bfde951bdf429f702351e943de6aaaaccc55abf5775b28d7795"} Sep 30 14:08:59 crc kubenswrapper[4840]: I0930 14:08:59.719055 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-8qmmm" podStartSLOduration=2.1058416 podStartE2EDuration="9.719031467s" podCreationTimestamp="2025-09-30 14:08:50 +0000 UTC" firstStartedPulling="2025-09-30 14:08:51.480085825 +0000 UTC m=+760.109172248" lastFinishedPulling="2025-09-30 14:08:59.093275692 +0000 UTC m=+767.722362115" observedRunningTime="2025-09-30 14:08:59.715429685 +0000 UTC m=+768.344516108" watchObservedRunningTime="2025-09-30 14:08:59.719031467 +0000 UTC m=+768.348117890" Sep 30 14:09:00 crc kubenswrapper[4840]: I0930 14:09:00.667061 4840 generic.go:334] "Generic (PLEG): container finished" podID="624a00a9-8fda-4e40-9df0-e8452ed89d28" containerID="47bc3385704360b3f27b5f9ccc65669b6f6c9f8f3300623e66447e9f3c13fed4" exitCode=0 Sep 30 14:09:00 crc kubenswrapper[4840]: I0930 14:09:00.667270 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sqx2j" event={"ID":"624a00a9-8fda-4e40-9df0-e8452ed89d28","Type":"ContainerDied","Data":"47bc3385704360b3f27b5f9ccc65669b6f6c9f8f3300623e66447e9f3c13fed4"} Sep 30 14:09:01 crc kubenswrapper[4840]: I0930 14:09:01.407760 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-5d688f5ffc-nv2dz" Sep 30 14:09:01 crc kubenswrapper[4840]: I0930 14:09:01.675042 4840 generic.go:334] "Generic (PLEG): container finished" podID="6480ac95-e05f-40d5-b6e2-934aa91f0eb8" containerID="7df38d56a2ab0f6933cbdfc13cd6676c4c2c5c135b718561276a44e315dd3d03" exitCode=0 Sep 30 14:09:01 crc kubenswrapper[4840]: I0930 14:09:01.675098 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jhpfb" event={"ID":"6480ac95-e05f-40d5-b6e2-934aa91f0eb8","Type":"ContainerDied","Data":"7df38d56a2ab0f6933cbdfc13cd6676c4c2c5c135b718561276a44e315dd3d03"} Sep 30 14:09:01 crc kubenswrapper[4840]: I0930 14:09:01.679834 4840 generic.go:334] "Generic (PLEG): container finished" podID="624a00a9-8fda-4e40-9df0-e8452ed89d28" containerID="b4b8dfbdd59c3ccb352aecb936af1a3464d4a4c17832d879da72409df54096ee" exitCode=0 Sep 30 14:09:01 crc kubenswrapper[4840]: I0930 14:09:01.679892 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sqx2j" event={"ID":"624a00a9-8fda-4e40-9df0-e8452ed89d28","Type":"ContainerDied","Data":"b4b8dfbdd59c3ccb352aecb936af1a3464d4a4c17832d879da72409df54096ee"} Sep 30 14:09:02 crc kubenswrapper[4840]: I0930 14:09:02.288263 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-hnv4n" Sep 30 14:09:02 crc kubenswrapper[4840]: I0930 14:09:02.687267 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sqx2j" event={"ID":"624a00a9-8fda-4e40-9df0-e8452ed89d28","Type":"ContainerStarted","Data":"695b68f1f4b69f6189bdf1abfdebf41d22ae8a173494e5a4f52a9d318efba61f"} Sep 30 14:09:03 crc kubenswrapper[4840]: I0930 14:09:03.696165 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sqx2j" event={"ID":"624a00a9-8fda-4e40-9df0-e8452ed89d28","Type":"ContainerStarted","Data":"17663c6ed447adfefb42c09c6f157b6a338f055802483ee5922da9c1ffa5be3d"} Sep 30 14:09:03 crc kubenswrapper[4840]: I0930 14:09:03.696206 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sqx2j" event={"ID":"624a00a9-8fda-4e40-9df0-e8452ed89d28","Type":"ContainerStarted","Data":"1aec5222859069fb4374e7a7ad5a1b2fff7e4ef91d32246f15b3200954d39ec2"} Sep 30 14:09:05 crc kubenswrapper[4840]: I0930 14:09:05.543021 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-r6wf2"] Sep 30 14:09:05 crc kubenswrapper[4840]: I0930 14:09:05.544007 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-r6wf2" Sep 30 14:09:05 crc kubenswrapper[4840]: I0930 14:09:05.545705 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Sep 30 14:09:05 crc kubenswrapper[4840]: I0930 14:09:05.546197 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Sep 30 14:09:05 crc kubenswrapper[4840]: I0930 14:09:05.559630 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-r6wf2"] Sep 30 14:09:05 crc kubenswrapper[4840]: I0930 14:09:05.656973 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbjk9\" (UniqueName: \"kubernetes.io/projected/07de7b8e-5063-46dc-b1c6-eefbf3eb539c-kube-api-access-qbjk9\") pod \"openstack-operator-index-r6wf2\" (UID: \"07de7b8e-5063-46dc-b1c6-eefbf3eb539c\") " pod="openstack-operators/openstack-operator-index-r6wf2" Sep 30 14:09:05 crc kubenswrapper[4840]: I0930 14:09:05.758679 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbjk9\" (UniqueName: \"kubernetes.io/projected/07de7b8e-5063-46dc-b1c6-eefbf3eb539c-kube-api-access-qbjk9\") pod \"openstack-operator-index-r6wf2\" (UID: \"07de7b8e-5063-46dc-b1c6-eefbf3eb539c\") " pod="openstack-operators/openstack-operator-index-r6wf2" Sep 30 14:09:05 crc kubenswrapper[4840]: I0930 14:09:05.777791 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbjk9\" (UniqueName: \"kubernetes.io/projected/07de7b8e-5063-46dc-b1c6-eefbf3eb539c-kube-api-access-qbjk9\") pod \"openstack-operator-index-r6wf2\" (UID: \"07de7b8e-5063-46dc-b1c6-eefbf3eb539c\") " pod="openstack-operators/openstack-operator-index-r6wf2" Sep 30 14:09:05 crc kubenswrapper[4840]: I0930 14:09:05.862339 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-r6wf2" Sep 30 14:09:06 crc kubenswrapper[4840]: I0930 14:09:06.756392 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-r6wf2"] Sep 30 14:09:07 crc kubenswrapper[4840]: I0930 14:09:07.730015 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jhpfb" event={"ID":"6480ac95-e05f-40d5-b6e2-934aa91f0eb8","Type":"ContainerStarted","Data":"d188f4afd487ca52d9c8cdcf85f99ebb3957b384be8376fc75f96b32312f69f3"} Sep 30 14:09:07 crc kubenswrapper[4840]: I0930 14:09:07.731861 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-r6wf2" event={"ID":"07de7b8e-5063-46dc-b1c6-eefbf3eb539c","Type":"ContainerStarted","Data":"1a11220532c27af665166f9dab52dd79346c7c7f5eb61d5b38138dbe1c8049f8"} Sep 30 14:09:07 crc kubenswrapper[4840]: I0930 14:09:07.734993 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sqx2j" event={"ID":"624a00a9-8fda-4e40-9df0-e8452ed89d28","Type":"ContainerStarted","Data":"566029d514fa6ca0557c884a802f9d328ca0a8fb887782612c5ca9a2f993ca18"} Sep 30 14:09:08 crc kubenswrapper[4840]: I0930 14:09:08.747333 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sqx2j" event={"ID":"624a00a9-8fda-4e40-9df0-e8452ed89d28","Type":"ContainerStarted","Data":"edb17a4efd3f833a6d30fbff57a2aa39c9586535f38f48503b399d981ecfbea8"} Sep 30 14:09:08 crc kubenswrapper[4840]: I0930 14:09:08.747625 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sqx2j" event={"ID":"624a00a9-8fda-4e40-9df0-e8452ed89d28","Type":"ContainerStarted","Data":"7cdaabeec61a681cbc27cecb549609a2db120f1004b0698c8664a10ae5919526"} Sep 30 14:09:08 crc kubenswrapper[4840]: I0930 14:09:08.747735 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-sqx2j" Sep 30 14:09:08 crc kubenswrapper[4840]: I0930 14:09:08.772981 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-sqx2j" podStartSLOduration=10.515169557 podStartE2EDuration="18.77296241s" podCreationTimestamp="2025-09-30 14:08:50 +0000 UTC" firstStartedPulling="2025-09-30 14:08:50.854606228 +0000 UTC m=+759.483692651" lastFinishedPulling="2025-09-30 14:08:59.112399081 +0000 UTC m=+767.741485504" observedRunningTime="2025-09-30 14:09:08.772316964 +0000 UTC m=+777.401403387" watchObservedRunningTime="2025-09-30 14:09:08.77296241 +0000 UTC m=+777.402048833" Sep 30 14:09:08 crc kubenswrapper[4840]: I0930 14:09:08.793572 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jhpfb" podStartSLOduration=6.108305336 podStartE2EDuration="12.793539066s" podCreationTimestamp="2025-09-30 14:08:56 +0000 UTC" firstStartedPulling="2025-09-30 14:08:59.657535485 +0000 UTC m=+768.286621918" lastFinishedPulling="2025-09-30 14:09:06.342769225 +0000 UTC m=+774.971855648" observedRunningTime="2025-09-30 14:09:08.791888374 +0000 UTC m=+777.420974797" watchObservedRunningTime="2025-09-30 14:09:08.793539066 +0000 UTC m=+777.422625489" Sep 30 14:09:08 crc kubenswrapper[4840]: I0930 14:09:08.921090 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-r6wf2"] Sep 30 14:09:09 crc kubenswrapper[4840]: I0930 14:09:09.732496 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-gl8hl"] Sep 30 14:09:09 crc kubenswrapper[4840]: I0930 14:09:09.733300 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-gl8hl" Sep 30 14:09:09 crc kubenswrapper[4840]: I0930 14:09:09.735453 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-lpjzv" Sep 30 14:09:09 crc kubenswrapper[4840]: I0930 14:09:09.739825 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-gl8hl"] Sep 30 14:09:09 crc kubenswrapper[4840]: I0930 14:09:09.814821 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9z47v\" (UniqueName: \"kubernetes.io/projected/a1690286-e0ea-4cf7-877b-d5e172dc365e-kube-api-access-9z47v\") pod \"openstack-operator-index-gl8hl\" (UID: \"a1690286-e0ea-4cf7-877b-d5e172dc365e\") " pod="openstack-operators/openstack-operator-index-gl8hl" Sep 30 14:09:09 crc kubenswrapper[4840]: I0930 14:09:09.916242 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9z47v\" (UniqueName: \"kubernetes.io/projected/a1690286-e0ea-4cf7-877b-d5e172dc365e-kube-api-access-9z47v\") pod \"openstack-operator-index-gl8hl\" (UID: \"a1690286-e0ea-4cf7-877b-d5e172dc365e\") " pod="openstack-operators/openstack-operator-index-gl8hl" Sep 30 14:09:09 crc kubenswrapper[4840]: I0930 14:09:09.934674 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9z47v\" (UniqueName: \"kubernetes.io/projected/a1690286-e0ea-4cf7-877b-d5e172dc365e-kube-api-access-9z47v\") pod \"openstack-operator-index-gl8hl\" (UID: \"a1690286-e0ea-4cf7-877b-d5e172dc365e\") " pod="openstack-operators/openstack-operator-index-gl8hl" Sep 30 14:09:10 crc kubenswrapper[4840]: I0930 14:09:10.047651 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-gl8hl" Sep 30 14:09:10 crc kubenswrapper[4840]: I0930 14:09:10.666234 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-sqx2j" Sep 30 14:09:10 crc kubenswrapper[4840]: I0930 14:09:10.707445 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-sqx2j" Sep 30 14:09:10 crc kubenswrapper[4840]: I0930 14:09:10.761298 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-r6wf2" event={"ID":"07de7b8e-5063-46dc-b1c6-eefbf3eb539c","Type":"ContainerStarted","Data":"72ad62135a96b38b96d67eb01a89e5394c25b892f09754d0ccfa30079caf6302"} Sep 30 14:09:10 crc kubenswrapper[4840]: I0930 14:09:10.761499 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-r6wf2" podUID="07de7b8e-5063-46dc-b1c6-eefbf3eb539c" containerName="registry-server" containerID="cri-o://72ad62135a96b38b96d67eb01a89e5394c25b892f09754d0ccfa30079caf6302" gracePeriod=2 Sep 30 14:09:10 crc kubenswrapper[4840]: I0930 14:09:10.776978 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-r6wf2" podStartSLOduration=1.932515769 podStartE2EDuration="5.77696341s" podCreationTimestamp="2025-09-30 14:09:05 +0000 UTC" firstStartedPulling="2025-09-30 14:09:06.763698384 +0000 UTC m=+775.392784817" lastFinishedPulling="2025-09-30 14:09:10.608146035 +0000 UTC m=+779.237232458" observedRunningTime="2025-09-30 14:09:10.775859922 +0000 UTC m=+779.404946345" watchObservedRunningTime="2025-09-30 14:09:10.77696341 +0000 UTC m=+779.406049823" Sep 30 14:09:10 crc kubenswrapper[4840]: I0930 14:09:10.938486 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-gl8hl"] Sep 30 14:09:10 crc kubenswrapper[4840]: W0930 14:09:10.943412 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda1690286_e0ea_4cf7_877b_d5e172dc365e.slice/crio-40d185c8b5f530adf3c94b1902f37ffe2a23b1ad76be0c5ed329666f47fa5c55 WatchSource:0}: Error finding container 40d185c8b5f530adf3c94b1902f37ffe2a23b1ad76be0c5ed329666f47fa5c55: Status 404 returned error can't find the container with id 40d185c8b5f530adf3c94b1902f37ffe2a23b1ad76be0c5ed329666f47fa5c55 Sep 30 14:09:11 crc kubenswrapper[4840]: I0930 14:09:11.181629 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-r6wf2_07de7b8e-5063-46dc-b1c6-eefbf3eb539c/registry-server/0.log" Sep 30 14:09:11 crc kubenswrapper[4840]: I0930 14:09:11.181946 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-r6wf2" Sep 30 14:09:11 crc kubenswrapper[4840]: I0930 14:09:11.256516 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-8qmmm" Sep 30 14:09:11 crc kubenswrapper[4840]: I0930 14:09:11.334692 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbjk9\" (UniqueName: \"kubernetes.io/projected/07de7b8e-5063-46dc-b1c6-eefbf3eb539c-kube-api-access-qbjk9\") pod \"07de7b8e-5063-46dc-b1c6-eefbf3eb539c\" (UID: \"07de7b8e-5063-46dc-b1c6-eefbf3eb539c\") " Sep 30 14:09:11 crc kubenswrapper[4840]: I0930 14:09:11.340421 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07de7b8e-5063-46dc-b1c6-eefbf3eb539c-kube-api-access-qbjk9" (OuterVolumeSpecName: "kube-api-access-qbjk9") pod "07de7b8e-5063-46dc-b1c6-eefbf3eb539c" (UID: "07de7b8e-5063-46dc-b1c6-eefbf3eb539c"). InnerVolumeSpecName "kube-api-access-qbjk9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:09:11 crc kubenswrapper[4840]: I0930 14:09:11.436221 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbjk9\" (UniqueName: \"kubernetes.io/projected/07de7b8e-5063-46dc-b1c6-eefbf3eb539c-kube-api-access-qbjk9\") on node \"crc\" DevicePath \"\"" Sep 30 14:09:11 crc kubenswrapper[4840]: I0930 14:09:11.769172 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-gl8hl" event={"ID":"a1690286-e0ea-4cf7-877b-d5e172dc365e","Type":"ContainerStarted","Data":"2c14fc0b8efb9b2b12a81f7e93cc9c72c26452e58d2d4cb2098ee5c8aeaffa4b"} Sep 30 14:09:11 crc kubenswrapper[4840]: I0930 14:09:11.769226 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-gl8hl" event={"ID":"a1690286-e0ea-4cf7-877b-d5e172dc365e","Type":"ContainerStarted","Data":"40d185c8b5f530adf3c94b1902f37ffe2a23b1ad76be0c5ed329666f47fa5c55"} Sep 30 14:09:11 crc kubenswrapper[4840]: I0930 14:09:11.771134 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-r6wf2_07de7b8e-5063-46dc-b1c6-eefbf3eb539c/registry-server/0.log" Sep 30 14:09:11 crc kubenswrapper[4840]: I0930 14:09:11.771194 4840 generic.go:334] "Generic (PLEG): container finished" podID="07de7b8e-5063-46dc-b1c6-eefbf3eb539c" containerID="72ad62135a96b38b96d67eb01a89e5394c25b892f09754d0ccfa30079caf6302" exitCode=2 Sep 30 14:09:11 crc kubenswrapper[4840]: I0930 14:09:11.771237 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-r6wf2" Sep 30 14:09:11 crc kubenswrapper[4840]: I0930 14:09:11.771290 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-r6wf2" event={"ID":"07de7b8e-5063-46dc-b1c6-eefbf3eb539c","Type":"ContainerDied","Data":"72ad62135a96b38b96d67eb01a89e5394c25b892f09754d0ccfa30079caf6302"} Sep 30 14:09:11 crc kubenswrapper[4840]: I0930 14:09:11.771377 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-r6wf2" event={"ID":"07de7b8e-5063-46dc-b1c6-eefbf3eb539c","Type":"ContainerDied","Data":"1a11220532c27af665166f9dab52dd79346c7c7f5eb61d5b38138dbe1c8049f8"} Sep 30 14:09:11 crc kubenswrapper[4840]: I0930 14:09:11.771408 4840 scope.go:117] "RemoveContainer" containerID="72ad62135a96b38b96d67eb01a89e5394c25b892f09754d0ccfa30079caf6302" Sep 30 14:09:11 crc kubenswrapper[4840]: I0930 14:09:11.776719 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-sqx2j" Sep 30 14:09:11 crc kubenswrapper[4840]: I0930 14:09:11.785633 4840 scope.go:117] "RemoveContainer" containerID="72ad62135a96b38b96d67eb01a89e5394c25b892f09754d0ccfa30079caf6302" Sep 30 14:09:11 crc kubenswrapper[4840]: E0930 14:09:11.786423 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72ad62135a96b38b96d67eb01a89e5394c25b892f09754d0ccfa30079caf6302\": container with ID starting with 72ad62135a96b38b96d67eb01a89e5394c25b892f09754d0ccfa30079caf6302 not found: ID does not exist" containerID="72ad62135a96b38b96d67eb01a89e5394c25b892f09754d0ccfa30079caf6302" Sep 30 14:09:11 crc kubenswrapper[4840]: I0930 14:09:11.786469 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72ad62135a96b38b96d67eb01a89e5394c25b892f09754d0ccfa30079caf6302"} err="failed to get container status \"72ad62135a96b38b96d67eb01a89e5394c25b892f09754d0ccfa30079caf6302\": rpc error: code = NotFound desc = could not find container \"72ad62135a96b38b96d67eb01a89e5394c25b892f09754d0ccfa30079caf6302\": container with ID starting with 72ad62135a96b38b96d67eb01a89e5394c25b892f09754d0ccfa30079caf6302 not found: ID does not exist" Sep 30 14:09:11 crc kubenswrapper[4840]: I0930 14:09:11.788449 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-gl8hl" podStartSLOduration=2.740636931 podStartE2EDuration="2.788433093s" podCreationTimestamp="2025-09-30 14:09:09 +0000 UTC" firstStartedPulling="2025-09-30 14:09:10.946982315 +0000 UTC m=+779.576068738" lastFinishedPulling="2025-09-30 14:09:10.994778477 +0000 UTC m=+779.623864900" observedRunningTime="2025-09-30 14:09:11.785233081 +0000 UTC m=+780.414319504" watchObservedRunningTime="2025-09-30 14:09:11.788433093 +0000 UTC m=+780.417519516" Sep 30 14:09:11 crc kubenswrapper[4840]: I0930 14:09:11.825131 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-r6wf2"] Sep 30 14:09:11 crc kubenswrapper[4840]: I0930 14:09:11.828386 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-r6wf2"] Sep 30 14:09:12 crc kubenswrapper[4840]: I0930 14:09:12.124096 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07de7b8e-5063-46dc-b1c6-eefbf3eb539c" path="/var/lib/kubelet/pods/07de7b8e-5063-46dc-b1c6-eefbf3eb539c/volumes" Sep 30 14:09:16 crc kubenswrapper[4840]: I0930 14:09:16.884027 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jhpfb" Sep 30 14:09:16 crc kubenswrapper[4840]: I0930 14:09:16.884514 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jhpfb" Sep 30 14:09:16 crc kubenswrapper[4840]: I0930 14:09:16.926477 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jhpfb" Sep 30 14:09:17 crc kubenswrapper[4840]: I0930 14:09:17.836950 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jhpfb" Sep 30 14:09:19 crc kubenswrapper[4840]: I0930 14:09:19.323228 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jhpfb"] Sep 30 14:09:19 crc kubenswrapper[4840]: I0930 14:09:19.825824 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jhpfb" podUID="6480ac95-e05f-40d5-b6e2-934aa91f0eb8" containerName="registry-server" containerID="cri-o://d188f4afd487ca52d9c8cdcf85f99ebb3957b384be8376fc75f96b32312f69f3" gracePeriod=2 Sep 30 14:09:20 crc kubenswrapper[4840]: I0930 14:09:20.047790 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-gl8hl" Sep 30 14:09:20 crc kubenswrapper[4840]: I0930 14:09:20.049086 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-gl8hl" Sep 30 14:09:20 crc kubenswrapper[4840]: I0930 14:09:20.082873 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-gl8hl" Sep 30 14:09:20 crc kubenswrapper[4840]: I0930 14:09:20.262235 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jhpfb" Sep 30 14:09:20 crc kubenswrapper[4840]: I0930 14:09:20.351255 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6480ac95-e05f-40d5-b6e2-934aa91f0eb8-utilities\") pod \"6480ac95-e05f-40d5-b6e2-934aa91f0eb8\" (UID: \"6480ac95-e05f-40d5-b6e2-934aa91f0eb8\") " Sep 30 14:09:20 crc kubenswrapper[4840]: I0930 14:09:20.351384 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6mtw\" (UniqueName: \"kubernetes.io/projected/6480ac95-e05f-40d5-b6e2-934aa91f0eb8-kube-api-access-r6mtw\") pod \"6480ac95-e05f-40d5-b6e2-934aa91f0eb8\" (UID: \"6480ac95-e05f-40d5-b6e2-934aa91f0eb8\") " Sep 30 14:09:20 crc kubenswrapper[4840]: I0930 14:09:20.351415 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6480ac95-e05f-40d5-b6e2-934aa91f0eb8-catalog-content\") pod \"6480ac95-e05f-40d5-b6e2-934aa91f0eb8\" (UID: \"6480ac95-e05f-40d5-b6e2-934aa91f0eb8\") " Sep 30 14:09:20 crc kubenswrapper[4840]: I0930 14:09:20.352236 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6480ac95-e05f-40d5-b6e2-934aa91f0eb8-utilities" (OuterVolumeSpecName: "utilities") pod "6480ac95-e05f-40d5-b6e2-934aa91f0eb8" (UID: "6480ac95-e05f-40d5-b6e2-934aa91f0eb8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:09:20 crc kubenswrapper[4840]: I0930 14:09:20.356224 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6480ac95-e05f-40d5-b6e2-934aa91f0eb8-kube-api-access-r6mtw" (OuterVolumeSpecName: "kube-api-access-r6mtw") pod "6480ac95-e05f-40d5-b6e2-934aa91f0eb8" (UID: "6480ac95-e05f-40d5-b6e2-934aa91f0eb8"). InnerVolumeSpecName "kube-api-access-r6mtw". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:09:20 crc kubenswrapper[4840]: I0930 14:09:20.391149 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6480ac95-e05f-40d5-b6e2-934aa91f0eb8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6480ac95-e05f-40d5-b6e2-934aa91f0eb8" (UID: "6480ac95-e05f-40d5-b6e2-934aa91f0eb8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:09:20 crc kubenswrapper[4840]: I0930 14:09:20.453418 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6mtw\" (UniqueName: \"kubernetes.io/projected/6480ac95-e05f-40d5-b6e2-934aa91f0eb8-kube-api-access-r6mtw\") on node \"crc\" DevicePath \"\"" Sep 30 14:09:20 crc kubenswrapper[4840]: I0930 14:09:20.453452 4840 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6480ac95-e05f-40d5-b6e2-934aa91f0eb8-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:09:20 crc kubenswrapper[4840]: I0930 14:09:20.453461 4840 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6480ac95-e05f-40d5-b6e2-934aa91f0eb8-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:09:20 crc kubenswrapper[4840]: I0930 14:09:20.833159 4840 generic.go:334] "Generic (PLEG): container finished" podID="6480ac95-e05f-40d5-b6e2-934aa91f0eb8" containerID="d188f4afd487ca52d9c8cdcf85f99ebb3957b384be8376fc75f96b32312f69f3" exitCode=0 Sep 30 14:09:20 crc kubenswrapper[4840]: I0930 14:09:20.833230 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jhpfb" event={"ID":"6480ac95-e05f-40d5-b6e2-934aa91f0eb8","Type":"ContainerDied","Data":"d188f4afd487ca52d9c8cdcf85f99ebb3957b384be8376fc75f96b32312f69f3"} Sep 30 14:09:20 crc kubenswrapper[4840]: I0930 14:09:20.833278 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jhpfb" event={"ID":"6480ac95-e05f-40d5-b6e2-934aa91f0eb8","Type":"ContainerDied","Data":"9d183a2457b3eabb33ee32715f1170f6e558f939df42a40b2c8f8415075c17f3"} Sep 30 14:09:20 crc kubenswrapper[4840]: I0930 14:09:20.833296 4840 scope.go:117] "RemoveContainer" containerID="d188f4afd487ca52d9c8cdcf85f99ebb3957b384be8376fc75f96b32312f69f3" Sep 30 14:09:20 crc kubenswrapper[4840]: I0930 14:09:20.833241 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jhpfb" Sep 30 14:09:20 crc kubenswrapper[4840]: I0930 14:09:20.847634 4840 scope.go:117] "RemoveContainer" containerID="7df38d56a2ab0f6933cbdfc13cd6676c4c2c5c135b718561276a44e315dd3d03" Sep 30 14:09:20 crc kubenswrapper[4840]: I0930 14:09:20.865765 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-gl8hl" Sep 30 14:09:20 crc kubenswrapper[4840]: I0930 14:09:20.867311 4840 scope.go:117] "RemoveContainer" containerID="392f8ef1ade10ace1c9b635fdb9932d87d576e38dc065e74c87893971f7efbe1" Sep 30 14:09:20 crc kubenswrapper[4840]: I0930 14:09:20.867427 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jhpfb"] Sep 30 14:09:20 crc kubenswrapper[4840]: I0930 14:09:20.870068 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jhpfb"] Sep 30 14:09:20 crc kubenswrapper[4840]: I0930 14:09:20.891925 4840 scope.go:117] "RemoveContainer" containerID="d188f4afd487ca52d9c8cdcf85f99ebb3957b384be8376fc75f96b32312f69f3" Sep 30 14:09:20 crc kubenswrapper[4840]: E0930 14:09:20.892659 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d188f4afd487ca52d9c8cdcf85f99ebb3957b384be8376fc75f96b32312f69f3\": container with ID starting with d188f4afd487ca52d9c8cdcf85f99ebb3957b384be8376fc75f96b32312f69f3 not found: ID does not exist" containerID="d188f4afd487ca52d9c8cdcf85f99ebb3957b384be8376fc75f96b32312f69f3" Sep 30 14:09:20 crc kubenswrapper[4840]: I0930 14:09:20.892694 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d188f4afd487ca52d9c8cdcf85f99ebb3957b384be8376fc75f96b32312f69f3"} err="failed to get container status \"d188f4afd487ca52d9c8cdcf85f99ebb3957b384be8376fc75f96b32312f69f3\": rpc error: code = NotFound desc = could not find container \"d188f4afd487ca52d9c8cdcf85f99ebb3957b384be8376fc75f96b32312f69f3\": container with ID starting with d188f4afd487ca52d9c8cdcf85f99ebb3957b384be8376fc75f96b32312f69f3 not found: ID does not exist" Sep 30 14:09:20 crc kubenswrapper[4840]: I0930 14:09:20.892717 4840 scope.go:117] "RemoveContainer" containerID="7df38d56a2ab0f6933cbdfc13cd6676c4c2c5c135b718561276a44e315dd3d03" Sep 30 14:09:20 crc kubenswrapper[4840]: E0930 14:09:20.893128 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7df38d56a2ab0f6933cbdfc13cd6676c4c2c5c135b718561276a44e315dd3d03\": container with ID starting with 7df38d56a2ab0f6933cbdfc13cd6676c4c2c5c135b718561276a44e315dd3d03 not found: ID does not exist" containerID="7df38d56a2ab0f6933cbdfc13cd6676c4c2c5c135b718561276a44e315dd3d03" Sep 30 14:09:20 crc kubenswrapper[4840]: I0930 14:09:20.893182 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7df38d56a2ab0f6933cbdfc13cd6676c4c2c5c135b718561276a44e315dd3d03"} err="failed to get container status \"7df38d56a2ab0f6933cbdfc13cd6676c4c2c5c135b718561276a44e315dd3d03\": rpc error: code = NotFound desc = could not find container \"7df38d56a2ab0f6933cbdfc13cd6676c4c2c5c135b718561276a44e315dd3d03\": container with ID starting with 7df38d56a2ab0f6933cbdfc13cd6676c4c2c5c135b718561276a44e315dd3d03 not found: ID does not exist" Sep 30 14:09:20 crc kubenswrapper[4840]: I0930 14:09:20.893214 4840 scope.go:117] "RemoveContainer" containerID="392f8ef1ade10ace1c9b635fdb9932d87d576e38dc065e74c87893971f7efbe1" Sep 30 14:09:20 crc kubenswrapper[4840]: E0930 14:09:20.893976 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"392f8ef1ade10ace1c9b635fdb9932d87d576e38dc065e74c87893971f7efbe1\": container with ID starting with 392f8ef1ade10ace1c9b635fdb9932d87d576e38dc065e74c87893971f7efbe1 not found: ID does not exist" containerID="392f8ef1ade10ace1c9b635fdb9932d87d576e38dc065e74c87893971f7efbe1" Sep 30 14:09:20 crc kubenswrapper[4840]: I0930 14:09:20.894006 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"392f8ef1ade10ace1c9b635fdb9932d87d576e38dc065e74c87893971f7efbe1"} err="failed to get container status \"392f8ef1ade10ace1c9b635fdb9932d87d576e38dc065e74c87893971f7efbe1\": rpc error: code = NotFound desc = could not find container \"392f8ef1ade10ace1c9b635fdb9932d87d576e38dc065e74c87893971f7efbe1\": container with ID starting with 392f8ef1ade10ace1c9b635fdb9932d87d576e38dc065e74c87893971f7efbe1 not found: ID does not exist" Sep 30 14:09:21 crc kubenswrapper[4840]: I0930 14:09:21.871496 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:09:21 crc kubenswrapper[4840]: I0930 14:09:21.871910 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:09:21 crc kubenswrapper[4840]: I0930 14:09:21.957155 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx"] Sep 30 14:09:21 crc kubenswrapper[4840]: E0930 14:09:21.957697 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6480ac95-e05f-40d5-b6e2-934aa91f0eb8" containerName="registry-server" Sep 30 14:09:21 crc kubenswrapper[4840]: I0930 14:09:21.957803 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="6480ac95-e05f-40d5-b6e2-934aa91f0eb8" containerName="registry-server" Sep 30 14:09:21 crc kubenswrapper[4840]: E0930 14:09:21.957900 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6480ac95-e05f-40d5-b6e2-934aa91f0eb8" containerName="extract-content" Sep 30 14:09:21 crc kubenswrapper[4840]: I0930 14:09:21.957982 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="6480ac95-e05f-40d5-b6e2-934aa91f0eb8" containerName="extract-content" Sep 30 14:09:21 crc kubenswrapper[4840]: E0930 14:09:21.958068 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07de7b8e-5063-46dc-b1c6-eefbf3eb539c" containerName="registry-server" Sep 30 14:09:21 crc kubenswrapper[4840]: I0930 14:09:21.958140 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="07de7b8e-5063-46dc-b1c6-eefbf3eb539c" containerName="registry-server" Sep 30 14:09:21 crc kubenswrapper[4840]: E0930 14:09:21.958227 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6480ac95-e05f-40d5-b6e2-934aa91f0eb8" containerName="extract-utilities" Sep 30 14:09:21 crc kubenswrapper[4840]: I0930 14:09:21.958321 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="6480ac95-e05f-40d5-b6e2-934aa91f0eb8" containerName="extract-utilities" Sep 30 14:09:21 crc kubenswrapper[4840]: I0930 14:09:21.958657 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="07de7b8e-5063-46dc-b1c6-eefbf3eb539c" containerName="registry-server" Sep 30 14:09:21 crc kubenswrapper[4840]: I0930 14:09:21.959010 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="6480ac95-e05f-40d5-b6e2-934aa91f0eb8" containerName="registry-server" Sep 30 14:09:21 crc kubenswrapper[4840]: I0930 14:09:21.960154 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx" Sep 30 14:09:21 crc kubenswrapper[4840]: I0930 14:09:21.962400 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-79d5j" Sep 30 14:09:21 crc kubenswrapper[4840]: I0930 14:09:21.964641 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx"] Sep 30 14:09:22 crc kubenswrapper[4840]: I0930 14:09:22.074213 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpns2\" (UniqueName: \"kubernetes.io/projected/181ca13b-195f-45ea-9894-3383d192d377-kube-api-access-vpns2\") pod \"06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx\" (UID: \"181ca13b-195f-45ea-9894-3383d192d377\") " pod="openstack-operators/06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx" Sep 30 14:09:22 crc kubenswrapper[4840]: I0930 14:09:22.074276 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/181ca13b-195f-45ea-9894-3383d192d377-bundle\") pod \"06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx\" (UID: \"181ca13b-195f-45ea-9894-3383d192d377\") " pod="openstack-operators/06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx" Sep 30 14:09:22 crc kubenswrapper[4840]: I0930 14:09:22.074318 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/181ca13b-195f-45ea-9894-3383d192d377-util\") pod \"06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx\" (UID: \"181ca13b-195f-45ea-9894-3383d192d377\") " pod="openstack-operators/06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx" Sep 30 14:09:22 crc kubenswrapper[4840]: I0930 14:09:22.124204 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6480ac95-e05f-40d5-b6e2-934aa91f0eb8" path="/var/lib/kubelet/pods/6480ac95-e05f-40d5-b6e2-934aa91f0eb8/volumes" Sep 30 14:09:22 crc kubenswrapper[4840]: I0930 14:09:22.175374 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpns2\" (UniqueName: \"kubernetes.io/projected/181ca13b-195f-45ea-9894-3383d192d377-kube-api-access-vpns2\") pod \"06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx\" (UID: \"181ca13b-195f-45ea-9894-3383d192d377\") " pod="openstack-operators/06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx" Sep 30 14:09:22 crc kubenswrapper[4840]: I0930 14:09:22.175422 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/181ca13b-195f-45ea-9894-3383d192d377-bundle\") pod \"06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx\" (UID: \"181ca13b-195f-45ea-9894-3383d192d377\") " pod="openstack-operators/06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx" Sep 30 14:09:22 crc kubenswrapper[4840]: I0930 14:09:22.175465 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/181ca13b-195f-45ea-9894-3383d192d377-util\") pod \"06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx\" (UID: \"181ca13b-195f-45ea-9894-3383d192d377\") " pod="openstack-operators/06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx" Sep 30 14:09:22 crc kubenswrapper[4840]: I0930 14:09:22.176017 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/181ca13b-195f-45ea-9894-3383d192d377-util\") pod \"06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx\" (UID: \"181ca13b-195f-45ea-9894-3383d192d377\") " pod="openstack-operators/06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx" Sep 30 14:09:22 crc kubenswrapper[4840]: I0930 14:09:22.176235 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/181ca13b-195f-45ea-9894-3383d192d377-bundle\") pod \"06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx\" (UID: \"181ca13b-195f-45ea-9894-3383d192d377\") " pod="openstack-operators/06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx" Sep 30 14:09:22 crc kubenswrapper[4840]: I0930 14:09:22.193990 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpns2\" (UniqueName: \"kubernetes.io/projected/181ca13b-195f-45ea-9894-3383d192d377-kube-api-access-vpns2\") pod \"06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx\" (UID: \"181ca13b-195f-45ea-9894-3383d192d377\") " pod="openstack-operators/06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx" Sep 30 14:09:22 crc kubenswrapper[4840]: I0930 14:09:22.274697 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx" Sep 30 14:09:22 crc kubenswrapper[4840]: I0930 14:09:22.734561 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx"] Sep 30 14:09:22 crc kubenswrapper[4840]: W0930 14:09:22.746846 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod181ca13b_195f_45ea_9894_3383d192d377.slice/crio-a180556e38b0d7abf9c3b16b30e10dec0220d27f1ae437858cec35cb4d48cf15 WatchSource:0}: Error finding container a180556e38b0d7abf9c3b16b30e10dec0220d27f1ae437858cec35cb4d48cf15: Status 404 returned error can't find the container with id a180556e38b0d7abf9c3b16b30e10dec0220d27f1ae437858cec35cb4d48cf15 Sep 30 14:09:22 crc kubenswrapper[4840]: I0930 14:09:22.847085 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx" event={"ID":"181ca13b-195f-45ea-9894-3383d192d377","Type":"ContainerStarted","Data":"a180556e38b0d7abf9c3b16b30e10dec0220d27f1ae437858cec35cb4d48cf15"} Sep 30 14:09:23 crc kubenswrapper[4840]: I0930 14:09:23.857750 4840 generic.go:334] "Generic (PLEG): container finished" podID="181ca13b-195f-45ea-9894-3383d192d377" containerID="5447d3a07633d0590da4b63d5611f7797d8cd53ef77357d76bbad14f7fb57328" exitCode=0 Sep 30 14:09:23 crc kubenswrapper[4840]: I0930 14:09:23.857870 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx" event={"ID":"181ca13b-195f-45ea-9894-3383d192d377","Type":"ContainerDied","Data":"5447d3a07633d0590da4b63d5611f7797d8cd53ef77357d76bbad14f7fb57328"} Sep 30 14:09:24 crc kubenswrapper[4840]: I0930 14:09:24.867699 4840 generic.go:334] "Generic (PLEG): container finished" podID="181ca13b-195f-45ea-9894-3383d192d377" containerID="abcc24b4deccec290592e8e93d21e37c336af86379784c62f97d858fe3d5884a" exitCode=0 Sep 30 14:09:24 crc kubenswrapper[4840]: I0930 14:09:24.867768 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx" event={"ID":"181ca13b-195f-45ea-9894-3383d192d377","Type":"ContainerDied","Data":"abcc24b4deccec290592e8e93d21e37c336af86379784c62f97d858fe3d5884a"} Sep 30 14:09:25 crc kubenswrapper[4840]: I0930 14:09:25.874616 4840 generic.go:334] "Generic (PLEG): container finished" podID="181ca13b-195f-45ea-9894-3383d192d377" containerID="b902b716b2f1d438f400242e0cf1dd582c6a8578aa6719a341cffdaaabddb4e7" exitCode=0 Sep 30 14:09:25 crc kubenswrapper[4840]: I0930 14:09:25.874830 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx" event={"ID":"181ca13b-195f-45ea-9894-3383d192d377","Type":"ContainerDied","Data":"b902b716b2f1d438f400242e0cf1dd582c6a8578aa6719a341cffdaaabddb4e7"} Sep 30 14:09:27 crc kubenswrapper[4840]: I0930 14:09:27.171065 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx" Sep 30 14:09:27 crc kubenswrapper[4840]: I0930 14:09:27.348487 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/181ca13b-195f-45ea-9894-3383d192d377-util\") pod \"181ca13b-195f-45ea-9894-3383d192d377\" (UID: \"181ca13b-195f-45ea-9894-3383d192d377\") " Sep 30 14:09:27 crc kubenswrapper[4840]: I0930 14:09:27.348905 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/181ca13b-195f-45ea-9894-3383d192d377-bundle\") pod \"181ca13b-195f-45ea-9894-3383d192d377\" (UID: \"181ca13b-195f-45ea-9894-3383d192d377\") " Sep 30 14:09:27 crc kubenswrapper[4840]: I0930 14:09:27.348997 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vpns2\" (UniqueName: \"kubernetes.io/projected/181ca13b-195f-45ea-9894-3383d192d377-kube-api-access-vpns2\") pod \"181ca13b-195f-45ea-9894-3383d192d377\" (UID: \"181ca13b-195f-45ea-9894-3383d192d377\") " Sep 30 14:09:27 crc kubenswrapper[4840]: I0930 14:09:27.349376 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/181ca13b-195f-45ea-9894-3383d192d377-bundle" (OuterVolumeSpecName: "bundle") pod "181ca13b-195f-45ea-9894-3383d192d377" (UID: "181ca13b-195f-45ea-9894-3383d192d377"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:09:27 crc kubenswrapper[4840]: I0930 14:09:27.354421 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/181ca13b-195f-45ea-9894-3383d192d377-kube-api-access-vpns2" (OuterVolumeSpecName: "kube-api-access-vpns2") pod "181ca13b-195f-45ea-9894-3383d192d377" (UID: "181ca13b-195f-45ea-9894-3383d192d377"). InnerVolumeSpecName "kube-api-access-vpns2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:09:27 crc kubenswrapper[4840]: I0930 14:09:27.363883 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/181ca13b-195f-45ea-9894-3383d192d377-util" (OuterVolumeSpecName: "util") pod "181ca13b-195f-45ea-9894-3383d192d377" (UID: "181ca13b-195f-45ea-9894-3383d192d377"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:09:27 crc kubenswrapper[4840]: I0930 14:09:27.451240 4840 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/181ca13b-195f-45ea-9894-3383d192d377-util\") on node \"crc\" DevicePath \"\"" Sep 30 14:09:27 crc kubenswrapper[4840]: I0930 14:09:27.451292 4840 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/181ca13b-195f-45ea-9894-3383d192d377-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:09:27 crc kubenswrapper[4840]: I0930 14:09:27.451310 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vpns2\" (UniqueName: \"kubernetes.io/projected/181ca13b-195f-45ea-9894-3383d192d377-kube-api-access-vpns2\") on node \"crc\" DevicePath \"\"" Sep 30 14:09:27 crc kubenswrapper[4840]: I0930 14:09:27.887053 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx" event={"ID":"181ca13b-195f-45ea-9894-3383d192d377","Type":"ContainerDied","Data":"a180556e38b0d7abf9c3b16b30e10dec0220d27f1ae437858cec35cb4d48cf15"} Sep 30 14:09:27 crc kubenswrapper[4840]: I0930 14:09:27.887100 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a180556e38b0d7abf9c3b16b30e10dec0220d27f1ae437858cec35cb4d48cf15" Sep 30 14:09:27 crc kubenswrapper[4840]: I0930 14:09:27.887115 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx" Sep 30 14:09:31 crc kubenswrapper[4840]: I0930 14:09:31.497388 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-84594b7f86-xwqzh"] Sep 30 14:09:31 crc kubenswrapper[4840]: E0930 14:09:31.498171 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="181ca13b-195f-45ea-9894-3383d192d377" containerName="util" Sep 30 14:09:31 crc kubenswrapper[4840]: I0930 14:09:31.498186 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="181ca13b-195f-45ea-9894-3383d192d377" containerName="util" Sep 30 14:09:31 crc kubenswrapper[4840]: E0930 14:09:31.498196 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="181ca13b-195f-45ea-9894-3383d192d377" containerName="extract" Sep 30 14:09:31 crc kubenswrapper[4840]: I0930 14:09:31.498204 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="181ca13b-195f-45ea-9894-3383d192d377" containerName="extract" Sep 30 14:09:31 crc kubenswrapper[4840]: E0930 14:09:31.498221 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="181ca13b-195f-45ea-9894-3383d192d377" containerName="pull" Sep 30 14:09:31 crc kubenswrapper[4840]: I0930 14:09:31.498229 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="181ca13b-195f-45ea-9894-3383d192d377" containerName="pull" Sep 30 14:09:31 crc kubenswrapper[4840]: I0930 14:09:31.498359 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="181ca13b-195f-45ea-9894-3383d192d377" containerName="extract" Sep 30 14:09:31 crc kubenswrapper[4840]: I0930 14:09:31.499104 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-84594b7f86-xwqzh" Sep 30 14:09:31 crc kubenswrapper[4840]: I0930 14:09:31.503428 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-4xhr2" Sep 30 14:09:31 crc kubenswrapper[4840]: I0930 14:09:31.517214 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-84594b7f86-xwqzh"] Sep 30 14:09:31 crc kubenswrapper[4840]: I0930 14:09:31.601032 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6f5x\" (UniqueName: \"kubernetes.io/projected/0b30b057-e591-411e-919b-5d1efac1a249-kube-api-access-p6f5x\") pod \"openstack-operator-controller-operator-84594b7f86-xwqzh\" (UID: \"0b30b057-e591-411e-919b-5d1efac1a249\") " pod="openstack-operators/openstack-operator-controller-operator-84594b7f86-xwqzh" Sep 30 14:09:31 crc kubenswrapper[4840]: I0930 14:09:31.702820 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6f5x\" (UniqueName: \"kubernetes.io/projected/0b30b057-e591-411e-919b-5d1efac1a249-kube-api-access-p6f5x\") pod \"openstack-operator-controller-operator-84594b7f86-xwqzh\" (UID: \"0b30b057-e591-411e-919b-5d1efac1a249\") " pod="openstack-operators/openstack-operator-controller-operator-84594b7f86-xwqzh" Sep 30 14:09:31 crc kubenswrapper[4840]: I0930 14:09:31.748694 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6f5x\" (UniqueName: \"kubernetes.io/projected/0b30b057-e591-411e-919b-5d1efac1a249-kube-api-access-p6f5x\") pod \"openstack-operator-controller-operator-84594b7f86-xwqzh\" (UID: \"0b30b057-e591-411e-919b-5d1efac1a249\") " pod="openstack-operators/openstack-operator-controller-operator-84594b7f86-xwqzh" Sep 30 14:09:31 crc kubenswrapper[4840]: I0930 14:09:31.817348 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-84594b7f86-xwqzh" Sep 30 14:09:32 crc kubenswrapper[4840]: I0930 14:09:32.267317 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-84594b7f86-xwqzh"] Sep 30 14:09:32 crc kubenswrapper[4840]: I0930 14:09:32.920049 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-84594b7f86-xwqzh" event={"ID":"0b30b057-e591-411e-919b-5d1efac1a249","Type":"ContainerStarted","Data":"0a944a06d39fa801b49df846affd8cbe1e325626215eb337eba17e32e84e8ccc"} Sep 30 14:09:35 crc kubenswrapper[4840]: I0930 14:09:35.940580 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-84594b7f86-xwqzh" event={"ID":"0b30b057-e591-411e-919b-5d1efac1a249","Type":"ContainerStarted","Data":"518a7c9daa672aef3beb70f2ff6987ebd0df4688f6e8ac0fed3b7d646a5e4e3a"} Sep 30 14:09:38 crc kubenswrapper[4840]: I0930 14:09:38.962194 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-84594b7f86-xwqzh" event={"ID":"0b30b057-e591-411e-919b-5d1efac1a249","Type":"ContainerStarted","Data":"d8f9ccf633592fe037682a83e03152a2d436b6b885d4e546e839d04774ad1fdf"} Sep 30 14:09:38 crc kubenswrapper[4840]: I0930 14:09:38.962508 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-84594b7f86-xwqzh" Sep 30 14:09:38 crc kubenswrapper[4840]: I0930 14:09:38.988301 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-84594b7f86-xwqzh" podStartSLOduration=2.375072095 podStartE2EDuration="7.988286335s" podCreationTimestamp="2025-09-30 14:09:31 +0000 UTC" firstStartedPulling="2025-09-30 14:09:32.276868376 +0000 UTC m=+800.905954799" lastFinishedPulling="2025-09-30 14:09:37.890082616 +0000 UTC m=+806.519169039" observedRunningTime="2025-09-30 14:09:38.985765141 +0000 UTC m=+807.614851574" watchObservedRunningTime="2025-09-30 14:09:38.988286335 +0000 UTC m=+807.617372758" Sep 30 14:09:39 crc kubenswrapper[4840]: I0930 14:09:39.532326 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jhbzc"] Sep 30 14:09:39 crc kubenswrapper[4840]: I0930 14:09:39.533637 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jhbzc" Sep 30 14:09:39 crc kubenswrapper[4840]: I0930 14:09:39.541647 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jhbzc"] Sep 30 14:09:39 crc kubenswrapper[4840]: I0930 14:09:39.611902 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80961b43-7573-4a66-879e-d151193a496c-utilities\") pod \"community-operators-jhbzc\" (UID: \"80961b43-7573-4a66-879e-d151193a496c\") " pod="openshift-marketplace/community-operators-jhbzc" Sep 30 14:09:39 crc kubenswrapper[4840]: I0930 14:09:39.611964 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qx4l4\" (UniqueName: \"kubernetes.io/projected/80961b43-7573-4a66-879e-d151193a496c-kube-api-access-qx4l4\") pod \"community-operators-jhbzc\" (UID: \"80961b43-7573-4a66-879e-d151193a496c\") " pod="openshift-marketplace/community-operators-jhbzc" Sep 30 14:09:39 crc kubenswrapper[4840]: I0930 14:09:39.612012 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80961b43-7573-4a66-879e-d151193a496c-catalog-content\") pod \"community-operators-jhbzc\" (UID: \"80961b43-7573-4a66-879e-d151193a496c\") " pod="openshift-marketplace/community-operators-jhbzc" Sep 30 14:09:39 crc kubenswrapper[4840]: I0930 14:09:39.714154 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80961b43-7573-4a66-879e-d151193a496c-utilities\") pod \"community-operators-jhbzc\" (UID: \"80961b43-7573-4a66-879e-d151193a496c\") " pod="openshift-marketplace/community-operators-jhbzc" Sep 30 14:09:39 crc kubenswrapper[4840]: I0930 14:09:39.714241 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qx4l4\" (UniqueName: \"kubernetes.io/projected/80961b43-7573-4a66-879e-d151193a496c-kube-api-access-qx4l4\") pod \"community-operators-jhbzc\" (UID: \"80961b43-7573-4a66-879e-d151193a496c\") " pod="openshift-marketplace/community-operators-jhbzc" Sep 30 14:09:39 crc kubenswrapper[4840]: I0930 14:09:39.714282 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80961b43-7573-4a66-879e-d151193a496c-catalog-content\") pod \"community-operators-jhbzc\" (UID: \"80961b43-7573-4a66-879e-d151193a496c\") " pod="openshift-marketplace/community-operators-jhbzc" Sep 30 14:09:39 crc kubenswrapper[4840]: I0930 14:09:39.714929 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80961b43-7573-4a66-879e-d151193a496c-utilities\") pod \"community-operators-jhbzc\" (UID: \"80961b43-7573-4a66-879e-d151193a496c\") " pod="openshift-marketplace/community-operators-jhbzc" Sep 30 14:09:39 crc kubenswrapper[4840]: I0930 14:09:39.715047 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80961b43-7573-4a66-879e-d151193a496c-catalog-content\") pod \"community-operators-jhbzc\" (UID: \"80961b43-7573-4a66-879e-d151193a496c\") " pod="openshift-marketplace/community-operators-jhbzc" Sep 30 14:09:39 crc kubenswrapper[4840]: I0930 14:09:39.733404 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qx4l4\" (UniqueName: \"kubernetes.io/projected/80961b43-7573-4a66-879e-d151193a496c-kube-api-access-qx4l4\") pod \"community-operators-jhbzc\" (UID: \"80961b43-7573-4a66-879e-d151193a496c\") " pod="openshift-marketplace/community-operators-jhbzc" Sep 30 14:09:39 crc kubenswrapper[4840]: I0930 14:09:39.851907 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jhbzc" Sep 30 14:09:40 crc kubenswrapper[4840]: I0930 14:09:40.314129 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jhbzc"] Sep 30 14:09:40 crc kubenswrapper[4840]: W0930 14:09:40.320779 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod80961b43_7573_4a66_879e_d151193a496c.slice/crio-008dcb850cc2182ea3984a5d067aadec3ebbbf785d5acfcafb0af7f40438f4c5 WatchSource:0}: Error finding container 008dcb850cc2182ea3984a5d067aadec3ebbbf785d5acfcafb0af7f40438f4c5: Status 404 returned error can't find the container with id 008dcb850cc2182ea3984a5d067aadec3ebbbf785d5acfcafb0af7f40438f4c5 Sep 30 14:09:40 crc kubenswrapper[4840]: I0930 14:09:40.974065 4840 generic.go:334] "Generic (PLEG): container finished" podID="80961b43-7573-4a66-879e-d151193a496c" containerID="ee7fca6b867f0504b77a9eef9682ad5fca9d39e84aa3c9289fd721363acfe5fc" exitCode=0 Sep 30 14:09:40 crc kubenswrapper[4840]: I0930 14:09:40.974172 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jhbzc" event={"ID":"80961b43-7573-4a66-879e-d151193a496c","Type":"ContainerDied","Data":"ee7fca6b867f0504b77a9eef9682ad5fca9d39e84aa3c9289fd721363acfe5fc"} Sep 30 14:09:40 crc kubenswrapper[4840]: I0930 14:09:40.974310 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jhbzc" event={"ID":"80961b43-7573-4a66-879e-d151193a496c","Type":"ContainerStarted","Data":"008dcb850cc2182ea3984a5d067aadec3ebbbf785d5acfcafb0af7f40438f4c5"} Sep 30 14:09:41 crc kubenswrapper[4840]: I0930 14:09:41.821406 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-84594b7f86-xwqzh" Sep 30 14:09:41 crc kubenswrapper[4840]: I0930 14:09:41.982347 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jhbzc" event={"ID":"80961b43-7573-4a66-879e-d151193a496c","Type":"ContainerStarted","Data":"a9914f3b78f05c49909efb7ea296dddc755176ab057160fcd30cf8e441578008"} Sep 30 14:09:42 crc kubenswrapper[4840]: I0930 14:09:42.990136 4840 generic.go:334] "Generic (PLEG): container finished" podID="80961b43-7573-4a66-879e-d151193a496c" containerID="a9914f3b78f05c49909efb7ea296dddc755176ab057160fcd30cf8e441578008" exitCode=0 Sep 30 14:09:42 crc kubenswrapper[4840]: I0930 14:09:42.990256 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jhbzc" event={"ID":"80961b43-7573-4a66-879e-d151193a496c","Type":"ContainerDied","Data":"a9914f3b78f05c49909efb7ea296dddc755176ab057160fcd30cf8e441578008"} Sep 30 14:09:43 crc kubenswrapper[4840]: I0930 14:09:43.998426 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jhbzc" event={"ID":"80961b43-7573-4a66-879e-d151193a496c","Type":"ContainerStarted","Data":"6aa77a3720695c61f1007270ada3db4de61a516d4e694dee1ba5238791713121"} Sep 30 14:09:44 crc kubenswrapper[4840]: I0930 14:09:44.020355 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jhbzc" podStartSLOduration=2.459851036 podStartE2EDuration="5.020333971s" podCreationTimestamp="2025-09-30 14:09:39 +0000 UTC" firstStartedPulling="2025-09-30 14:09:40.975430385 +0000 UTC m=+809.604516808" lastFinishedPulling="2025-09-30 14:09:43.53591331 +0000 UTC m=+812.164999743" observedRunningTime="2025-09-30 14:09:44.014071781 +0000 UTC m=+812.643158224" watchObservedRunningTime="2025-09-30 14:09:44.020333971 +0000 UTC m=+812.649420414" Sep 30 14:09:49 crc kubenswrapper[4840]: I0930 14:09:49.852342 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jhbzc" Sep 30 14:09:49 crc kubenswrapper[4840]: I0930 14:09:49.852950 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jhbzc" Sep 30 14:09:49 crc kubenswrapper[4840]: I0930 14:09:49.890183 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jhbzc" Sep 30 14:09:50 crc kubenswrapper[4840]: I0930 14:09:50.076421 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jhbzc" Sep 30 14:09:50 crc kubenswrapper[4840]: I0930 14:09:50.132696 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jhbzc"] Sep 30 14:09:51 crc kubenswrapper[4840]: I0930 14:09:51.871401 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:09:51 crc kubenswrapper[4840]: I0930 14:09:51.871736 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:09:52 crc kubenswrapper[4840]: I0930 14:09:52.041013 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jhbzc" podUID="80961b43-7573-4a66-879e-d151193a496c" containerName="registry-server" containerID="cri-o://6aa77a3720695c61f1007270ada3db4de61a516d4e694dee1ba5238791713121" gracePeriod=2 Sep 30 14:09:52 crc kubenswrapper[4840]: I0930 14:09:52.428881 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jhbzc" Sep 30 14:09:52 crc kubenswrapper[4840]: I0930 14:09:52.577980 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qx4l4\" (UniqueName: \"kubernetes.io/projected/80961b43-7573-4a66-879e-d151193a496c-kube-api-access-qx4l4\") pod \"80961b43-7573-4a66-879e-d151193a496c\" (UID: \"80961b43-7573-4a66-879e-d151193a496c\") " Sep 30 14:09:52 crc kubenswrapper[4840]: I0930 14:09:52.578061 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80961b43-7573-4a66-879e-d151193a496c-catalog-content\") pod \"80961b43-7573-4a66-879e-d151193a496c\" (UID: \"80961b43-7573-4a66-879e-d151193a496c\") " Sep 30 14:09:52 crc kubenswrapper[4840]: I0930 14:09:52.578116 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80961b43-7573-4a66-879e-d151193a496c-utilities\") pod \"80961b43-7573-4a66-879e-d151193a496c\" (UID: \"80961b43-7573-4a66-879e-d151193a496c\") " Sep 30 14:09:52 crc kubenswrapper[4840]: I0930 14:09:52.579251 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80961b43-7573-4a66-879e-d151193a496c-utilities" (OuterVolumeSpecName: "utilities") pod "80961b43-7573-4a66-879e-d151193a496c" (UID: "80961b43-7573-4a66-879e-d151193a496c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:09:52 crc kubenswrapper[4840]: I0930 14:09:52.583469 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80961b43-7573-4a66-879e-d151193a496c-kube-api-access-qx4l4" (OuterVolumeSpecName: "kube-api-access-qx4l4") pod "80961b43-7573-4a66-879e-d151193a496c" (UID: "80961b43-7573-4a66-879e-d151193a496c"). InnerVolumeSpecName "kube-api-access-qx4l4". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:09:52 crc kubenswrapper[4840]: I0930 14:09:52.679832 4840 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80961b43-7573-4a66-879e-d151193a496c-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:09:52 crc kubenswrapper[4840]: I0930 14:09:52.679867 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qx4l4\" (UniqueName: \"kubernetes.io/projected/80961b43-7573-4a66-879e-d151193a496c-kube-api-access-qx4l4\") on node \"crc\" DevicePath \"\"" Sep 30 14:09:53 crc kubenswrapper[4840]: I0930 14:09:53.048205 4840 generic.go:334] "Generic (PLEG): container finished" podID="80961b43-7573-4a66-879e-d151193a496c" containerID="6aa77a3720695c61f1007270ada3db4de61a516d4e694dee1ba5238791713121" exitCode=0 Sep 30 14:09:53 crc kubenswrapper[4840]: I0930 14:09:53.048272 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jhbzc" Sep 30 14:09:53 crc kubenswrapper[4840]: I0930 14:09:53.048829 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jhbzc" event={"ID":"80961b43-7573-4a66-879e-d151193a496c","Type":"ContainerDied","Data":"6aa77a3720695c61f1007270ada3db4de61a516d4e694dee1ba5238791713121"} Sep 30 14:09:53 crc kubenswrapper[4840]: I0930 14:09:53.048887 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jhbzc" event={"ID":"80961b43-7573-4a66-879e-d151193a496c","Type":"ContainerDied","Data":"008dcb850cc2182ea3984a5d067aadec3ebbbf785d5acfcafb0af7f40438f4c5"} Sep 30 14:09:53 crc kubenswrapper[4840]: I0930 14:09:53.048905 4840 scope.go:117] "RemoveContainer" containerID="6aa77a3720695c61f1007270ada3db4de61a516d4e694dee1ba5238791713121" Sep 30 14:09:53 crc kubenswrapper[4840]: I0930 14:09:53.066266 4840 scope.go:117] "RemoveContainer" containerID="a9914f3b78f05c49909efb7ea296dddc755176ab057160fcd30cf8e441578008" Sep 30 14:09:53 crc kubenswrapper[4840]: I0930 14:09:53.081636 4840 scope.go:117] "RemoveContainer" containerID="ee7fca6b867f0504b77a9eef9682ad5fca9d39e84aa3c9289fd721363acfe5fc" Sep 30 14:09:53 crc kubenswrapper[4840]: I0930 14:09:53.097998 4840 scope.go:117] "RemoveContainer" containerID="6aa77a3720695c61f1007270ada3db4de61a516d4e694dee1ba5238791713121" Sep 30 14:09:53 crc kubenswrapper[4840]: E0930 14:09:53.098377 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6aa77a3720695c61f1007270ada3db4de61a516d4e694dee1ba5238791713121\": container with ID starting with 6aa77a3720695c61f1007270ada3db4de61a516d4e694dee1ba5238791713121 not found: ID does not exist" containerID="6aa77a3720695c61f1007270ada3db4de61a516d4e694dee1ba5238791713121" Sep 30 14:09:53 crc kubenswrapper[4840]: I0930 14:09:53.098418 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6aa77a3720695c61f1007270ada3db4de61a516d4e694dee1ba5238791713121"} err="failed to get container status \"6aa77a3720695c61f1007270ada3db4de61a516d4e694dee1ba5238791713121\": rpc error: code = NotFound desc = could not find container \"6aa77a3720695c61f1007270ada3db4de61a516d4e694dee1ba5238791713121\": container with ID starting with 6aa77a3720695c61f1007270ada3db4de61a516d4e694dee1ba5238791713121 not found: ID does not exist" Sep 30 14:09:53 crc kubenswrapper[4840]: I0930 14:09:53.098443 4840 scope.go:117] "RemoveContainer" containerID="a9914f3b78f05c49909efb7ea296dddc755176ab057160fcd30cf8e441578008" Sep 30 14:09:53 crc kubenswrapper[4840]: E0930 14:09:53.098783 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9914f3b78f05c49909efb7ea296dddc755176ab057160fcd30cf8e441578008\": container with ID starting with a9914f3b78f05c49909efb7ea296dddc755176ab057160fcd30cf8e441578008 not found: ID does not exist" containerID="a9914f3b78f05c49909efb7ea296dddc755176ab057160fcd30cf8e441578008" Sep 30 14:09:53 crc kubenswrapper[4840]: I0930 14:09:53.098897 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9914f3b78f05c49909efb7ea296dddc755176ab057160fcd30cf8e441578008"} err="failed to get container status \"a9914f3b78f05c49909efb7ea296dddc755176ab057160fcd30cf8e441578008\": rpc error: code = NotFound desc = could not find container \"a9914f3b78f05c49909efb7ea296dddc755176ab057160fcd30cf8e441578008\": container with ID starting with a9914f3b78f05c49909efb7ea296dddc755176ab057160fcd30cf8e441578008 not found: ID does not exist" Sep 30 14:09:53 crc kubenswrapper[4840]: I0930 14:09:53.098991 4840 scope.go:117] "RemoveContainer" containerID="ee7fca6b867f0504b77a9eef9682ad5fca9d39e84aa3c9289fd721363acfe5fc" Sep 30 14:09:53 crc kubenswrapper[4840]: E0930 14:09:53.099450 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee7fca6b867f0504b77a9eef9682ad5fca9d39e84aa3c9289fd721363acfe5fc\": container with ID starting with ee7fca6b867f0504b77a9eef9682ad5fca9d39e84aa3c9289fd721363acfe5fc not found: ID does not exist" containerID="ee7fca6b867f0504b77a9eef9682ad5fca9d39e84aa3c9289fd721363acfe5fc" Sep 30 14:09:53 crc kubenswrapper[4840]: I0930 14:09:53.099523 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee7fca6b867f0504b77a9eef9682ad5fca9d39e84aa3c9289fd721363acfe5fc"} err="failed to get container status \"ee7fca6b867f0504b77a9eef9682ad5fca9d39e84aa3c9289fd721363acfe5fc\": rpc error: code = NotFound desc = could not find container \"ee7fca6b867f0504b77a9eef9682ad5fca9d39e84aa3c9289fd721363acfe5fc\": container with ID starting with ee7fca6b867f0504b77a9eef9682ad5fca9d39e84aa3c9289fd721363acfe5fc not found: ID does not exist" Sep 30 14:09:53 crc kubenswrapper[4840]: I0930 14:09:53.190101 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80961b43-7573-4a66-879e-d151193a496c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "80961b43-7573-4a66-879e-d151193a496c" (UID: "80961b43-7573-4a66-879e-d151193a496c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:09:53 crc kubenswrapper[4840]: I0930 14:09:53.287692 4840 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80961b43-7573-4a66-879e-d151193a496c-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:09:53 crc kubenswrapper[4840]: I0930 14:09:53.376884 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jhbzc"] Sep 30 14:09:53 crc kubenswrapper[4840]: I0930 14:09:53.383861 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jhbzc"] Sep 30 14:09:54 crc kubenswrapper[4840]: I0930 14:09:54.130841 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80961b43-7573-4a66-879e-d151193a496c" path="/var/lib/kubelet/pods/80961b43-7573-4a66-879e-d151193a496c/volumes" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.589298 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-rnsbn"] Sep 30 14:10:08 crc kubenswrapper[4840]: E0930 14:10:08.590333 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80961b43-7573-4a66-879e-d151193a496c" containerName="registry-server" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.590357 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="80961b43-7573-4a66-879e-d151193a496c" containerName="registry-server" Sep 30 14:10:08 crc kubenswrapper[4840]: E0930 14:10:08.590392 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80961b43-7573-4a66-879e-d151193a496c" containerName="extract-utilities" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.590404 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="80961b43-7573-4a66-879e-d151193a496c" containerName="extract-utilities" Sep 30 14:10:08 crc kubenswrapper[4840]: E0930 14:10:08.590421 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80961b43-7573-4a66-879e-d151193a496c" containerName="extract-content" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.590434 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="80961b43-7573-4a66-879e-d151193a496c" containerName="extract-content" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.590668 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="80961b43-7573-4a66-879e-d151193a496c" containerName="registry-server" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.591629 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-rnsbn" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.600373 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-rnsbn"] Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.600997 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-smg6p"] Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.604300 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-xpw69" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.606793 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-smg6p" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.608340 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-vjv99" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.618008 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-smg6p"] Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.652059 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-6nn29"] Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.653165 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-6nn29" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.659354 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-kmtrk" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.664275 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-6nn29"] Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.687231 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdf69\" (UniqueName: \"kubernetes.io/projected/3ffb0dd0-3dd8-4f37-8ead-4add746a1fc6-kube-api-access-hdf69\") pod \"cinder-operator-controller-manager-644bddb6d8-smg6p\" (UID: \"3ffb0dd0-3dd8-4f37-8ead-4add746a1fc6\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-smg6p" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.687368 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86t6z\" (UniqueName: \"kubernetes.io/projected/0f4e505b-4892-4c92-9ec4-0bdeb5d4c873-kube-api-access-86t6z\") pod \"barbican-operator-controller-manager-6ff8b75857-rnsbn\" (UID: \"0f4e505b-4892-4c92-9ec4-0bdeb5d4c873\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-rnsbn" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.687389 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-sq5hd"] Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.688542 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-sq5hd" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.696294 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-tdmpf" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.698668 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-jwlj9"] Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.699786 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-jwlj9" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.702304 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-rrxng" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.718595 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-d9jgv"] Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.725128 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-d9jgv" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.732515 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-d2qlb" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.749014 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-sq5hd"] Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.788206 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wd5rd\" (UniqueName: \"kubernetes.io/projected/64c5ad91-f113-46e3-ac87-36521d32f4f0-kube-api-access-wd5rd\") pod \"glance-operator-controller-manager-84958c4d49-sq5hd\" (UID: \"64c5ad91-f113-46e3-ac87-36521d32f4f0\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-sq5hd" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.788253 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9cw2\" (UniqueName: \"kubernetes.io/projected/63d94c36-0a55-472c-bbe8-27195105584e-kube-api-access-g9cw2\") pod \"horizon-operator-controller-manager-9f4696d94-d9jgv\" (UID: \"63d94c36-0a55-472c-bbe8-27195105584e\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-d9jgv" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.788296 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mcjd\" (UniqueName: \"kubernetes.io/projected/395025e2-9d12-4172-bb73-b4fb67e0111c-kube-api-access-8mcjd\") pod \"designate-operator-controller-manager-84f4f7b77b-6nn29\" (UID: \"395025e2-9d12-4172-bb73-b4fb67e0111c\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-6nn29" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.788380 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86t6z\" (UniqueName: \"kubernetes.io/projected/0f4e505b-4892-4c92-9ec4-0bdeb5d4c873-kube-api-access-86t6z\") pod \"barbican-operator-controller-manager-6ff8b75857-rnsbn\" (UID: \"0f4e505b-4892-4c92-9ec4-0bdeb5d4c873\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-rnsbn" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.788412 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdf69\" (UniqueName: \"kubernetes.io/projected/3ffb0dd0-3dd8-4f37-8ead-4add746a1fc6-kube-api-access-hdf69\") pod \"cinder-operator-controller-manager-644bddb6d8-smg6p\" (UID: \"3ffb0dd0-3dd8-4f37-8ead-4add746a1fc6\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-smg6p" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.788764 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptdb7\" (UniqueName: \"kubernetes.io/projected/5daf5662-8848-4fd9-a6af-e091fe303d89-kube-api-access-ptdb7\") pod \"heat-operator-controller-manager-5d889d78cf-jwlj9\" (UID: \"5daf5662-8848-4fd9-a6af-e091fe303d89\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-jwlj9" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.815596 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-7d857cc749-xtqwd"] Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.816503 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-7975b88857-fg7c7"] Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.816782 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-xtqwd" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.817268 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-fg7c7" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.821633 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-d9jgv"] Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.827613 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-dhjdb" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.827782 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.827870 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-vsqbb" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.832412 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86t6z\" (UniqueName: \"kubernetes.io/projected/0f4e505b-4892-4c92-9ec4-0bdeb5d4c873-kube-api-access-86t6z\") pod \"barbican-operator-controller-manager-6ff8b75857-rnsbn\" (UID: \"0f4e505b-4892-4c92-9ec4-0bdeb5d4c873\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-rnsbn" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.834141 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdf69\" (UniqueName: \"kubernetes.io/projected/3ffb0dd0-3dd8-4f37-8ead-4add746a1fc6-kube-api-access-hdf69\") pod \"cinder-operator-controller-manager-644bddb6d8-smg6p\" (UID: \"3ffb0dd0-3dd8-4f37-8ead-4add746a1fc6\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-smg6p" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.852894 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5bd55b4bff-rjpbm"] Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.853861 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-rjpbm" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.860970 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-g9f8s" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.876935 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-7975b88857-fg7c7"] Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.891914 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-d2hgd"] Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.892978 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-d2hgd" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.896863 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9cw2\" (UniqueName: \"kubernetes.io/projected/63d94c36-0a55-472c-bbe8-27195105584e-kube-api-access-g9cw2\") pod \"horizon-operator-controller-manager-9f4696d94-d9jgv\" (UID: \"63d94c36-0a55-472c-bbe8-27195105584e\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-d9jgv" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.896903 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gp2ng\" (UniqueName: \"kubernetes.io/projected/494a382c-6fec-4711-8a90-7f691c847e86-kube-api-access-gp2ng\") pod \"ironic-operator-controller-manager-7975b88857-fg7c7\" (UID: \"494a382c-6fec-4711-8a90-7f691c847e86\") " pod="openstack-operators/ironic-operator-controller-manager-7975b88857-fg7c7" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.896925 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mcjd\" (UniqueName: \"kubernetes.io/projected/395025e2-9d12-4172-bb73-b4fb67e0111c-kube-api-access-8mcjd\") pod \"designate-operator-controller-manager-84f4f7b77b-6nn29\" (UID: \"395025e2-9d12-4172-bb73-b4fb67e0111c\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-6nn29" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.896956 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/807f2091-02fa-41e0-943a-b30f4202c476-cert\") pod \"infra-operator-controller-manager-7d857cc749-xtqwd\" (UID: \"807f2091-02fa-41e0-943a-b30f4202c476\") " pod="openstack-operators/infra-operator-controller-manager-7d857cc749-xtqwd" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.897003 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptdb7\" (UniqueName: \"kubernetes.io/projected/5daf5662-8848-4fd9-a6af-e091fe303d89-kube-api-access-ptdb7\") pod \"heat-operator-controller-manager-5d889d78cf-jwlj9\" (UID: \"5daf5662-8848-4fd9-a6af-e091fe303d89\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-jwlj9" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.897029 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgd86\" (UniqueName: \"kubernetes.io/projected/807f2091-02fa-41e0-943a-b30f4202c476-kube-api-access-pgd86\") pod \"infra-operator-controller-manager-7d857cc749-xtqwd\" (UID: \"807f2091-02fa-41e0-943a-b30f4202c476\") " pod="openstack-operators/infra-operator-controller-manager-7d857cc749-xtqwd" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.897070 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wd5rd\" (UniqueName: \"kubernetes.io/projected/64c5ad91-f113-46e3-ac87-36521d32f4f0-kube-api-access-wd5rd\") pod \"glance-operator-controller-manager-84958c4d49-sq5hd\" (UID: \"64c5ad91-f113-46e3-ac87-36521d32f4f0\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-sq5hd" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.897070 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-w64n2" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.903460 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-jwlj9"] Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.919712 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-rnsbn" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.937320 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-smg6p" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.937543 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mcjd\" (UniqueName: \"kubernetes.io/projected/395025e2-9d12-4172-bb73-b4fb67e0111c-kube-api-access-8mcjd\") pod \"designate-operator-controller-manager-84f4f7b77b-6nn29\" (UID: \"395025e2-9d12-4172-bb73-b4fb67e0111c\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-6nn29" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.938259 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9cw2\" (UniqueName: \"kubernetes.io/projected/63d94c36-0a55-472c-bbe8-27195105584e-kube-api-access-g9cw2\") pod \"horizon-operator-controller-manager-9f4696d94-d9jgv\" (UID: \"63d94c36-0a55-472c-bbe8-27195105584e\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-d9jgv" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.956629 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-xzmg8"] Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.957650 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-88c7-xzmg8" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.958621 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wd5rd\" (UniqueName: \"kubernetes.io/projected/64c5ad91-f113-46e3-ac87-36521d32f4f0-kube-api-access-wd5rd\") pod \"glance-operator-controller-manager-84958c4d49-sq5hd\" (UID: \"64c5ad91-f113-46e3-ac87-36521d32f4f0\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-sq5hd" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.960753 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5bd55b4bff-rjpbm"] Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.990680 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptdb7\" (UniqueName: \"kubernetes.io/projected/5daf5662-8848-4fd9-a6af-e091fe303d89-kube-api-access-ptdb7\") pod \"heat-operator-controller-manager-5d889d78cf-jwlj9\" (UID: \"5daf5662-8848-4fd9-a6af-e091fe303d89\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-jwlj9" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.991373 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-vdspl" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.991511 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-d2hgd"] Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.991758 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-6nn29" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.998122 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgd86\" (UniqueName: \"kubernetes.io/projected/807f2091-02fa-41e0-943a-b30f4202c476-kube-api-access-pgd86\") pod \"infra-operator-controller-manager-7d857cc749-xtqwd\" (UID: \"807f2091-02fa-41e0-943a-b30f4202c476\") " pod="openstack-operators/infra-operator-controller-manager-7d857cc749-xtqwd" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.998188 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gp2ng\" (UniqueName: \"kubernetes.io/projected/494a382c-6fec-4711-8a90-7f691c847e86-kube-api-access-gp2ng\") pod \"ironic-operator-controller-manager-7975b88857-fg7c7\" (UID: \"494a382c-6fec-4711-8a90-7f691c847e86\") " pod="openstack-operators/ironic-operator-controller-manager-7975b88857-fg7c7" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.998216 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4x42g\" (UniqueName: \"kubernetes.io/projected/b927784b-0b44-4bf1-9751-3aafcc24367b-kube-api-access-4x42g\") pod \"keystone-operator-controller-manager-5bd55b4bff-rjpbm\" (UID: \"b927784b-0b44-4bf1-9751-3aafcc24367b\") " pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-rjpbm" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.998241 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/807f2091-02fa-41e0-943a-b30f4202c476-cert\") pod \"infra-operator-controller-manager-7d857cc749-xtqwd\" (UID: \"807f2091-02fa-41e0-943a-b30f4202c476\") " pod="openstack-operators/infra-operator-controller-manager-7d857cc749-xtqwd" Sep 30 14:10:08 crc kubenswrapper[4840]: I0930 14:10:08.998288 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9fws\" (UniqueName: \"kubernetes.io/projected/eb939f27-ce94-408d-b673-4c25f96dd059-kube-api-access-b9fws\") pod \"manila-operator-controller-manager-6d68dbc695-d2hgd\" (UID: \"eb939f27-ce94-408d-b673-4c25f96dd059\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-d2hgd" Sep 30 14:10:08 crc kubenswrapper[4840]: E0930 14:10:08.998471 4840 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Sep 30 14:10:08 crc kubenswrapper[4840]: E0930 14:10:08.998516 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/807f2091-02fa-41e0-943a-b30f4202c476-cert podName:807f2091-02fa-41e0-943a-b30f4202c476 nodeName:}" failed. No retries permitted until 2025-09-30 14:10:09.498501593 +0000 UTC m=+838.127588016 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/807f2091-02fa-41e0-943a-b30f4202c476-cert") pod "infra-operator-controller-manager-7d857cc749-xtqwd" (UID: "807f2091-02fa-41e0-943a-b30f4202c476") : secret "infra-operator-webhook-server-cert" not found Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.004998 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-xzmg8"] Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.005291 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-sq5hd" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.015802 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-7d857cc749-xtqwd"] Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.022051 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-c7c776c96-x4mx4"] Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.035545 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-jwlj9" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.036884 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-x4mx4" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.043902 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-g452f" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.044324 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64d7b59854-qsq84"] Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.045312 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-t2mx4"] Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.046084 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-t2mx4" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.046472 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-qsq84" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.047194 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-d9jgv" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.049985 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64d7b59854-qsq84"] Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.056049 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-scdm2" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.056197 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgd86\" (UniqueName: \"kubernetes.io/projected/807f2091-02fa-41e0-943a-b30f4202c476-kube-api-access-pgd86\") pod \"infra-operator-controller-manager-7d857cc749-xtqwd\" (UID: \"807f2091-02fa-41e0-943a-b30f4202c476\") " pod="openstack-operators/infra-operator-controller-manager-7d857cc749-xtqwd" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.056291 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-kn9rn" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.065101 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-c7c776c96-x4mx4"] Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.077940 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gp2ng\" (UniqueName: \"kubernetes.io/projected/494a382c-6fec-4711-8a90-7f691c847e86-kube-api-access-gp2ng\") pod \"ironic-operator-controller-manager-7975b88857-fg7c7\" (UID: \"494a382c-6fec-4711-8a90-7f691c847e86\") " pod="openstack-operators/ironic-operator-controller-manager-7975b88857-fg7c7" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.090108 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-p6msl"] Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.091068 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-t2mx4"] Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.091146 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-p6msl" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.101366 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-9t25m"] Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.101932 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9fws\" (UniqueName: \"kubernetes.io/projected/eb939f27-ce94-408d-b673-4c25f96dd059-kube-api-access-b9fws\") pod \"manila-operator-controller-manager-6d68dbc695-d2hgd\" (UID: \"eb939f27-ce94-408d-b673-4c25f96dd059\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-d2hgd" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.101988 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wbp9\" (UniqueName: \"kubernetes.io/projected/3f4d112b-e6b8-455a-9f66-e8ee1bb81ac9-kube-api-access-9wbp9\") pod \"nova-operator-controller-manager-c7c776c96-x4mx4\" (UID: \"3f4d112b-e6b8-455a-9f66-e8ee1bb81ac9\") " pod="openstack-operators/nova-operator-controller-manager-c7c776c96-x4mx4" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.102025 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k879m\" (UniqueName: \"kubernetes.io/projected/3d5da746-e5f3-4bed-9c77-37a78da8f1a6-kube-api-access-k879m\") pod \"neutron-operator-controller-manager-64d7b59854-qsq84\" (UID: \"3d5da746-e5f3-4bed-9c77-37a78da8f1a6\") " pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-qsq84" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.102048 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcqwx\" (UniqueName: \"kubernetes.io/projected/a76ab4ec-b180-473f-a597-641725526b58-kube-api-access-lcqwx\") pod \"octavia-operator-controller-manager-76fcc6dc7c-t2mx4\" (UID: \"a76ab4ec-b180-473f-a597-641725526b58\") " pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-t2mx4" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.102065 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4hxt\" (UniqueName: \"kubernetes.io/projected/385365f0-9e2c-4c89-b95b-164a473ff0c3-kube-api-access-w4hxt\") pod \"mariadb-operator-controller-manager-88c7-xzmg8\" (UID: \"385365f0-9e2c-4c89-b95b-164a473ff0c3\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-xzmg8" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.102085 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4x42g\" (UniqueName: \"kubernetes.io/projected/b927784b-0b44-4bf1-9751-3aafcc24367b-kube-api-access-4x42g\") pod \"keystone-operator-controller-manager-5bd55b4bff-rjpbm\" (UID: \"b927784b-0b44-4bf1-9751-3aafcc24367b\") " pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-rjpbm" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.102296 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-p6msl"] Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.102371 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-9t25m" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.106379 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.106720 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-vhl5r" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.106938 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-b2hrw" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.107083 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-tqjvl"] Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.108344 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-tqjvl" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.120746 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-9t25m"] Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.121287 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-6f49z" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.137752 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-tqjvl"] Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.145221 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-bc7dc7bd9-4mpb8"] Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.149697 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-4mpb8" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.152445 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-78974455b9-97rs4"] Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.164239 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-k9m6s" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.164277 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4x42g\" (UniqueName: \"kubernetes.io/projected/b927784b-0b44-4bf1-9751-3aafcc24367b-kube-api-access-4x42g\") pod \"keystone-operator-controller-manager-5bd55b4bff-rjpbm\" (UID: \"b927784b-0b44-4bf1-9751-3aafcc24367b\") " pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-rjpbm" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.164744 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9fws\" (UniqueName: \"kubernetes.io/projected/eb939f27-ce94-408d-b673-4c25f96dd059-kube-api-access-b9fws\") pod \"manila-operator-controller-manager-6d68dbc695-d2hgd\" (UID: \"eb939f27-ce94-408d-b673-4c25f96dd059\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-d2hgd" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.172313 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-78974455b9-97rs4" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.179365 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-b9s45" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.207788 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-fg7c7" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.209302 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6s8n\" (UniqueName: \"kubernetes.io/projected/3ed332eb-b604-4667-90f4-541d5a9115c2-kube-api-access-q6s8n\") pod \"placement-operator-controller-manager-589c58c6c-tqjvl\" (UID: \"3ed332eb-b604-4667-90f4-541d5a9115c2\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-tqjvl" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.209391 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/62475ec3-b883-4d16-bb31-6ee393062da2-cert\") pod \"openstack-baremetal-operator-controller-manager-6d776955-p6msl\" (UID: \"62475ec3-b883-4d16-bb31-6ee393062da2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-p6msl" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.209420 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nhpk\" (UniqueName: \"kubernetes.io/projected/62475ec3-b883-4d16-bb31-6ee393062da2-kube-api-access-6nhpk\") pod \"openstack-baremetal-operator-controller-manager-6d776955-p6msl\" (UID: \"62475ec3-b883-4d16-bb31-6ee393062da2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-p6msl" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.209457 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tgdf\" (UniqueName: \"kubernetes.io/projected/2b8703c9-f691-4c3a-9941-8bcd36aaf948-kube-api-access-6tgdf\") pod \"swift-operator-controller-manager-bc7dc7bd9-4mpb8\" (UID: \"2b8703c9-f691-4c3a-9941-8bcd36aaf948\") " pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-4mpb8" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.209517 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bksm\" (UniqueName: \"kubernetes.io/projected/dc75c0e8-691e-4108-b9c7-e4233d0baeb6-kube-api-access-2bksm\") pod \"ovn-operator-controller-manager-9976ff44c-9t25m\" (UID: \"dc75c0e8-691e-4108-b9c7-e4233d0baeb6\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-9t25m" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.209612 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wbp9\" (UniqueName: \"kubernetes.io/projected/3f4d112b-e6b8-455a-9f66-e8ee1bb81ac9-kube-api-access-9wbp9\") pod \"nova-operator-controller-manager-c7c776c96-x4mx4\" (UID: \"3f4d112b-e6b8-455a-9f66-e8ee1bb81ac9\") " pod="openstack-operators/nova-operator-controller-manager-c7c776c96-x4mx4" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.209663 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k879m\" (UniqueName: \"kubernetes.io/projected/3d5da746-e5f3-4bed-9c77-37a78da8f1a6-kube-api-access-k879m\") pod \"neutron-operator-controller-manager-64d7b59854-qsq84\" (UID: \"3d5da746-e5f3-4bed-9c77-37a78da8f1a6\") " pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-qsq84" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.209719 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcqwx\" (UniqueName: \"kubernetes.io/projected/a76ab4ec-b180-473f-a597-641725526b58-kube-api-access-lcqwx\") pod \"octavia-operator-controller-manager-76fcc6dc7c-t2mx4\" (UID: \"a76ab4ec-b180-473f-a597-641725526b58\") " pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-t2mx4" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.209743 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4hxt\" (UniqueName: \"kubernetes.io/projected/385365f0-9e2c-4c89-b95b-164a473ff0c3-kube-api-access-w4hxt\") pod \"mariadb-operator-controller-manager-88c7-xzmg8\" (UID: \"385365f0-9e2c-4c89-b95b-164a473ff0c3\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-xzmg8" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.220690 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-rjpbm" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.239627 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4hxt\" (UniqueName: \"kubernetes.io/projected/385365f0-9e2c-4c89-b95b-164a473ff0c3-kube-api-access-w4hxt\") pod \"mariadb-operator-controller-manager-88c7-xzmg8\" (UID: \"385365f0-9e2c-4c89-b95b-164a473ff0c3\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-xzmg8" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.247720 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcqwx\" (UniqueName: \"kubernetes.io/projected/a76ab4ec-b180-473f-a597-641725526b58-kube-api-access-lcqwx\") pod \"octavia-operator-controller-manager-76fcc6dc7c-t2mx4\" (UID: \"a76ab4ec-b180-473f-a597-641725526b58\") " pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-t2mx4" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.262213 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wbp9\" (UniqueName: \"kubernetes.io/projected/3f4d112b-e6b8-455a-9f66-e8ee1bb81ac9-kube-api-access-9wbp9\") pod \"nova-operator-controller-manager-c7c776c96-x4mx4\" (UID: \"3f4d112b-e6b8-455a-9f66-e8ee1bb81ac9\") " pod="openstack-operators/nova-operator-controller-manager-c7c776c96-x4mx4" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.262265 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-78974455b9-97rs4"] Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.265149 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k879m\" (UniqueName: \"kubernetes.io/projected/3d5da746-e5f3-4bed-9c77-37a78da8f1a6-kube-api-access-k879m\") pod \"neutron-operator-controller-manager-64d7b59854-qsq84\" (UID: \"3d5da746-e5f3-4bed-9c77-37a78da8f1a6\") " pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-qsq84" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.267832 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-bc7dc7bd9-4mpb8"] Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.277872 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-f66b554c6-9npm5"] Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.279167 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-f66b554c6-9npm5" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.286130 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-d5nkp" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.292115 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-f66b554c6-9npm5"] Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.311334 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6s8n\" (UniqueName: \"kubernetes.io/projected/3ed332eb-b604-4667-90f4-541d5a9115c2-kube-api-access-q6s8n\") pod \"placement-operator-controller-manager-589c58c6c-tqjvl\" (UID: \"3ed332eb-b604-4667-90f4-541d5a9115c2\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-tqjvl" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.311392 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/62475ec3-b883-4d16-bb31-6ee393062da2-cert\") pod \"openstack-baremetal-operator-controller-manager-6d776955-p6msl\" (UID: \"62475ec3-b883-4d16-bb31-6ee393062da2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-p6msl" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.311415 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nhpk\" (UniqueName: \"kubernetes.io/projected/62475ec3-b883-4d16-bb31-6ee393062da2-kube-api-access-6nhpk\") pod \"openstack-baremetal-operator-controller-manager-6d776955-p6msl\" (UID: \"62475ec3-b883-4d16-bb31-6ee393062da2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-p6msl" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.311437 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tgdf\" (UniqueName: \"kubernetes.io/projected/2b8703c9-f691-4c3a-9941-8bcd36aaf948-kube-api-access-6tgdf\") pod \"swift-operator-controller-manager-bc7dc7bd9-4mpb8\" (UID: \"2b8703c9-f691-4c3a-9941-8bcd36aaf948\") " pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-4mpb8" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.311458 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bksm\" (UniqueName: \"kubernetes.io/projected/dc75c0e8-691e-4108-b9c7-e4233d0baeb6-kube-api-access-2bksm\") pod \"ovn-operator-controller-manager-9976ff44c-9t25m\" (UID: \"dc75c0e8-691e-4108-b9c7-e4233d0baeb6\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-9t25m" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.311506 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bj2jb\" (UniqueName: \"kubernetes.io/projected/47248483-6556-491b-b65f-01cad9f1d1fc-kube-api-access-bj2jb\") pod \"telemetry-operator-controller-manager-78974455b9-97rs4\" (UID: \"47248483-6556-491b-b65f-01cad9f1d1fc\") " pod="openstack-operators/telemetry-operator-controller-manager-78974455b9-97rs4" Sep 30 14:10:09 crc kubenswrapper[4840]: E0930 14:10:09.311779 4840 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Sep 30 14:10:09 crc kubenswrapper[4840]: E0930 14:10:09.311873 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/62475ec3-b883-4d16-bb31-6ee393062da2-cert podName:62475ec3-b883-4d16-bb31-6ee393062da2 nodeName:}" failed. No retries permitted until 2025-09-30 14:10:09.811847802 +0000 UTC m=+838.440934275 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/62475ec3-b883-4d16-bb31-6ee393062da2-cert") pod "openstack-baremetal-operator-controller-manager-6d776955-p6msl" (UID: "62475ec3-b883-4d16-bb31-6ee393062da2") : secret "openstack-baremetal-operator-webhook-server-cert" not found Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.318442 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-76669f99c-rn2qq"] Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.319604 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-rn2qq" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.322362 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-phc2t" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.338158 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tgdf\" (UniqueName: \"kubernetes.io/projected/2b8703c9-f691-4c3a-9941-8bcd36aaf948-kube-api-access-6tgdf\") pod \"swift-operator-controller-manager-bc7dc7bd9-4mpb8\" (UID: \"2b8703c9-f691-4c3a-9941-8bcd36aaf948\") " pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-4mpb8" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.340864 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bksm\" (UniqueName: \"kubernetes.io/projected/dc75c0e8-691e-4108-b9c7-e4233d0baeb6-kube-api-access-2bksm\") pod \"ovn-operator-controller-manager-9976ff44c-9t25m\" (UID: \"dc75c0e8-691e-4108-b9c7-e4233d0baeb6\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-9t25m" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.344991 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nhpk\" (UniqueName: \"kubernetes.io/projected/62475ec3-b883-4d16-bb31-6ee393062da2-kube-api-access-6nhpk\") pod \"openstack-baremetal-operator-controller-manager-6d776955-p6msl\" (UID: \"62475ec3-b883-4d16-bb31-6ee393062da2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-p6msl" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.345695 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6s8n\" (UniqueName: \"kubernetes.io/projected/3ed332eb-b604-4667-90f4-541d5a9115c2-kube-api-access-q6s8n\") pod \"placement-operator-controller-manager-589c58c6c-tqjvl\" (UID: \"3ed332eb-b604-4667-90f4-541d5a9115c2\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-tqjvl" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.354620 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-4mpb8" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.361783 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-d2hgd" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.371639 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-76669f99c-rn2qq"] Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.382056 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-88c7-xzmg8" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.418412 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bj2jb\" (UniqueName: \"kubernetes.io/projected/47248483-6556-491b-b65f-01cad9f1d1fc-kube-api-access-bj2jb\") pod \"telemetry-operator-controller-manager-78974455b9-97rs4\" (UID: \"47248483-6556-491b-b65f-01cad9f1d1fc\") " pod="openstack-operators/telemetry-operator-controller-manager-78974455b9-97rs4" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.418480 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fv6j5\" (UniqueName: \"kubernetes.io/projected/e1cd519e-b5c3-48f2-b426-d0f4e912683b-kube-api-access-fv6j5\") pod \"watcher-operator-controller-manager-76669f99c-rn2qq\" (UID: \"e1cd519e-b5c3-48f2-b426-d0f4e912683b\") " pod="openstack-operators/watcher-operator-controller-manager-76669f99c-rn2qq" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.418592 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v98vj\" (UniqueName: \"kubernetes.io/projected/3ba6da78-9606-4e1c-bc55-6cbe2e052464-kube-api-access-v98vj\") pod \"test-operator-controller-manager-f66b554c6-9npm5\" (UID: \"3ba6da78-9606-4e1c-bc55-6cbe2e052464\") " pod="openstack-operators/test-operator-controller-manager-f66b554c6-9npm5" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.428236 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-x4mx4" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.444478 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bj2jb\" (UniqueName: \"kubernetes.io/projected/47248483-6556-491b-b65f-01cad9f1d1fc-kube-api-access-bj2jb\") pod \"telemetry-operator-controller-manager-78974455b9-97rs4\" (UID: \"47248483-6556-491b-b65f-01cad9f1d1fc\") " pod="openstack-operators/telemetry-operator-controller-manager-78974455b9-97rs4" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.473893 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-t2mx4" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.487831 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-568796b78d-bvbdn"] Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.490609 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-568796b78d-bvbdn" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.493427 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-vpmrs" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.493691 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.519448 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/807f2091-02fa-41e0-943a-b30f4202c476-cert\") pod \"infra-operator-controller-manager-7d857cc749-xtqwd\" (UID: \"807f2091-02fa-41e0-943a-b30f4202c476\") " pod="openstack-operators/infra-operator-controller-manager-7d857cc749-xtqwd" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.519508 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v98vj\" (UniqueName: \"kubernetes.io/projected/3ba6da78-9606-4e1c-bc55-6cbe2e052464-kube-api-access-v98vj\") pod \"test-operator-controller-manager-f66b554c6-9npm5\" (UID: \"3ba6da78-9606-4e1c-bc55-6cbe2e052464\") " pod="openstack-operators/test-operator-controller-manager-f66b554c6-9npm5" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.519961 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fv6j5\" (UniqueName: \"kubernetes.io/projected/e1cd519e-b5c3-48f2-b426-d0f4e912683b-kube-api-access-fv6j5\") pod \"watcher-operator-controller-manager-76669f99c-rn2qq\" (UID: \"e1cd519e-b5c3-48f2-b426-d0f4e912683b\") " pod="openstack-operators/watcher-operator-controller-manager-76669f99c-rn2qq" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.522613 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-568796b78d-bvbdn"] Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.533853 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-qsq84" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.534960 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/807f2091-02fa-41e0-943a-b30f4202c476-cert\") pod \"infra-operator-controller-manager-7d857cc749-xtqwd\" (UID: \"807f2091-02fa-41e0-943a-b30f4202c476\") " pod="openstack-operators/infra-operator-controller-manager-7d857cc749-xtqwd" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.542992 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v98vj\" (UniqueName: \"kubernetes.io/projected/3ba6da78-9606-4e1c-bc55-6cbe2e052464-kube-api-access-v98vj\") pod \"test-operator-controller-manager-f66b554c6-9npm5\" (UID: \"3ba6da78-9606-4e1c-bc55-6cbe2e052464\") " pod="openstack-operators/test-operator-controller-manager-f66b554c6-9npm5" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.555248 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-smjg7"] Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.556606 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-smjg7" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.562496 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-fp8jp" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.563064 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-smjg7"] Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.568783 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fv6j5\" (UniqueName: \"kubernetes.io/projected/e1cd519e-b5c3-48f2-b426-d0f4e912683b-kube-api-access-fv6j5\") pod \"watcher-operator-controller-manager-76669f99c-rn2qq\" (UID: \"e1cd519e-b5c3-48f2-b426-d0f4e912683b\") " pod="openstack-operators/watcher-operator-controller-manager-76669f99c-rn2qq" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.592690 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-9t25m" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.601878 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-rnsbn"] Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.613003 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-tqjvl" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.621286 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99j8v\" (UniqueName: \"kubernetes.io/projected/aa2cf241-ec41-4a94-b452-653b3439a073-kube-api-access-99j8v\") pod \"rabbitmq-cluster-operator-manager-79d8469568-smjg7\" (UID: \"aa2cf241-ec41-4a94-b452-653b3439a073\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-smjg7" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.621378 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e61dcaab-7705-4857-bdd9-adeb07d9e6f2-cert\") pod \"openstack-operator-controller-manager-568796b78d-bvbdn\" (UID: \"e61dcaab-7705-4857-bdd9-adeb07d9e6f2\") " pod="openstack-operators/openstack-operator-controller-manager-568796b78d-bvbdn" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.621423 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gt26z\" (UniqueName: \"kubernetes.io/projected/e61dcaab-7705-4857-bdd9-adeb07d9e6f2-kube-api-access-gt26z\") pod \"openstack-operator-controller-manager-568796b78d-bvbdn\" (UID: \"e61dcaab-7705-4857-bdd9-adeb07d9e6f2\") " pod="openstack-operators/openstack-operator-controller-manager-568796b78d-bvbdn" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.664186 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-78974455b9-97rs4" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.696479 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-f66b554c6-9npm5" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.700389 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-smg6p"] Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.716994 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-rn2qq" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.723525 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99j8v\" (UniqueName: \"kubernetes.io/projected/aa2cf241-ec41-4a94-b452-653b3439a073-kube-api-access-99j8v\") pod \"rabbitmq-cluster-operator-manager-79d8469568-smjg7\" (UID: \"aa2cf241-ec41-4a94-b452-653b3439a073\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-smjg7" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.723632 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e61dcaab-7705-4857-bdd9-adeb07d9e6f2-cert\") pod \"openstack-operator-controller-manager-568796b78d-bvbdn\" (UID: \"e61dcaab-7705-4857-bdd9-adeb07d9e6f2\") " pod="openstack-operators/openstack-operator-controller-manager-568796b78d-bvbdn" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.723675 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gt26z\" (UniqueName: \"kubernetes.io/projected/e61dcaab-7705-4857-bdd9-adeb07d9e6f2-kube-api-access-gt26z\") pod \"openstack-operator-controller-manager-568796b78d-bvbdn\" (UID: \"e61dcaab-7705-4857-bdd9-adeb07d9e6f2\") " pod="openstack-operators/openstack-operator-controller-manager-568796b78d-bvbdn" Sep 30 14:10:09 crc kubenswrapper[4840]: E0930 14:10:09.724156 4840 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Sep 30 14:10:09 crc kubenswrapper[4840]: E0930 14:10:09.724194 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e61dcaab-7705-4857-bdd9-adeb07d9e6f2-cert podName:e61dcaab-7705-4857-bdd9-adeb07d9e6f2 nodeName:}" failed. No retries permitted until 2025-09-30 14:10:10.224180581 +0000 UTC m=+838.853267004 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e61dcaab-7705-4857-bdd9-adeb07d9e6f2-cert") pod "openstack-operator-controller-manager-568796b78d-bvbdn" (UID: "e61dcaab-7705-4857-bdd9-adeb07d9e6f2") : secret "webhook-server-cert" not found Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.750897 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gt26z\" (UniqueName: \"kubernetes.io/projected/e61dcaab-7705-4857-bdd9-adeb07d9e6f2-kube-api-access-gt26z\") pod \"openstack-operator-controller-manager-568796b78d-bvbdn\" (UID: \"e61dcaab-7705-4857-bdd9-adeb07d9e6f2\") " pod="openstack-operators/openstack-operator-controller-manager-568796b78d-bvbdn" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.757452 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99j8v\" (UniqueName: \"kubernetes.io/projected/aa2cf241-ec41-4a94-b452-653b3439a073-kube-api-access-99j8v\") pod \"rabbitmq-cluster-operator-manager-79d8469568-smjg7\" (UID: \"aa2cf241-ec41-4a94-b452-653b3439a073\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-smjg7" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.792898 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-xtqwd" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.825441 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/62475ec3-b883-4d16-bb31-6ee393062da2-cert\") pod \"openstack-baremetal-operator-controller-manager-6d776955-p6msl\" (UID: \"62475ec3-b883-4d16-bb31-6ee393062da2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-p6msl" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.828693 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/62475ec3-b883-4d16-bb31-6ee393062da2-cert\") pod \"openstack-baremetal-operator-controller-manager-6d776955-p6msl\" (UID: \"62475ec3-b883-4d16-bb31-6ee393062da2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-p6msl" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.865829 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-p6msl" Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.886415 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-6nn29"] Sep 30 14:10:09 crc kubenswrapper[4840]: W0930 14:10:09.952358 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod395025e2_9d12_4172_bb73_b4fb67e0111c.slice/crio-7828760eb5741d5bd42e3dc0514de2aa33ef4fa78e438efa13c4ff3b58d518a9 WatchSource:0}: Error finding container 7828760eb5741d5bd42e3dc0514de2aa33ef4fa78e438efa13c4ff3b58d518a9: Status 404 returned error can't find the container with id 7828760eb5741d5bd42e3dc0514de2aa33ef4fa78e438efa13c4ff3b58d518a9 Sep 30 14:10:09 crc kubenswrapper[4840]: I0930 14:10:09.957029 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-smjg7" Sep 30 14:10:10 crc kubenswrapper[4840]: I0930 14:10:10.181512 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-rnsbn" event={"ID":"0f4e505b-4892-4c92-9ec4-0bdeb5d4c873","Type":"ContainerStarted","Data":"587bafbe46f5b42b11b049fb0d5ed4ccd78d6588a851714a1ca0a363d8fd1a75"} Sep 30 14:10:10 crc kubenswrapper[4840]: I0930 14:10:10.193159 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-smg6p" event={"ID":"3ffb0dd0-3dd8-4f37-8ead-4add746a1fc6","Type":"ContainerStarted","Data":"fc421e8fd7ad2107d7807c4788c09474b9b471e34d3aacb98645c525ef5f766d"} Sep 30 14:10:10 crc kubenswrapper[4840]: I0930 14:10:10.200084 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-6nn29" event={"ID":"395025e2-9d12-4172-bb73-b4fb67e0111c","Type":"ContainerStarted","Data":"7828760eb5741d5bd42e3dc0514de2aa33ef4fa78e438efa13c4ff3b58d518a9"} Sep 30 14:10:10 crc kubenswrapper[4840]: I0930 14:10:10.216677 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5bd55b4bff-rjpbm"] Sep 30 14:10:10 crc kubenswrapper[4840]: I0930 14:10:10.226466 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-sq5hd"] Sep 30 14:10:10 crc kubenswrapper[4840]: I0930 14:10:10.232338 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e61dcaab-7705-4857-bdd9-adeb07d9e6f2-cert\") pod \"openstack-operator-controller-manager-568796b78d-bvbdn\" (UID: \"e61dcaab-7705-4857-bdd9-adeb07d9e6f2\") " pod="openstack-operators/openstack-operator-controller-manager-568796b78d-bvbdn" Sep 30 14:10:10 crc kubenswrapper[4840]: I0930 14:10:10.235110 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-7975b88857-fg7c7"] Sep 30 14:10:10 crc kubenswrapper[4840]: I0930 14:10:10.244070 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e61dcaab-7705-4857-bdd9-adeb07d9e6f2-cert\") pod \"openstack-operator-controller-manager-568796b78d-bvbdn\" (UID: \"e61dcaab-7705-4857-bdd9-adeb07d9e6f2\") " pod="openstack-operators/openstack-operator-controller-manager-568796b78d-bvbdn" Sep 30 14:10:10 crc kubenswrapper[4840]: I0930 14:10:10.248336 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-jwlj9"] Sep 30 14:10:10 crc kubenswrapper[4840]: I0930 14:10:10.276832 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-bc7dc7bd9-4mpb8"] Sep 30 14:10:10 crc kubenswrapper[4840]: I0930 14:10:10.293373 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-xzmg8"] Sep 30 14:10:10 crc kubenswrapper[4840]: I0930 14:10:10.298475 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-d9jgv"] Sep 30 14:10:10 crc kubenswrapper[4840]: I0930 14:10:10.304693 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-d2hgd"] Sep 30 14:10:10 crc kubenswrapper[4840]: I0930 14:10:10.448937 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-9t25m"] Sep 30 14:10:10 crc kubenswrapper[4840]: W0930 14:10:10.454412 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddc75c0e8_691e_4108_b9c7_e4233d0baeb6.slice/crio-8e45575138ac2e33ec4b2a54d01152d1c484b24152856490a41f89a269f5bd31 WatchSource:0}: Error finding container 8e45575138ac2e33ec4b2a54d01152d1c484b24152856490a41f89a269f5bd31: Status 404 returned error can't find the container with id 8e45575138ac2e33ec4b2a54d01152d1c484b24152856490a41f89a269f5bd31 Sep 30 14:10:10 crc kubenswrapper[4840]: I0930 14:10:10.455802 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-568796b78d-bvbdn" Sep 30 14:10:10 crc kubenswrapper[4840]: W0930 14:10:10.456312 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda76ab4ec_b180_473f_a597_641725526b58.slice/crio-d90d64e6cd7098806ffbcdc8324c9b26488804248bde891b7f5122f4c231a204 WatchSource:0}: Error finding container d90d64e6cd7098806ffbcdc8324c9b26488804248bde891b7f5122f4c231a204: Status 404 returned error can't find the container with id d90d64e6cd7098806ffbcdc8324c9b26488804248bde891b7f5122f4c231a204 Sep 30 14:10:10 crc kubenswrapper[4840]: I0930 14:10:10.456425 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-c7c776c96-x4mx4"] Sep 30 14:10:10 crc kubenswrapper[4840]: W0930 14:10:10.457436 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3ed332eb_b604_4667_90f4_541d5a9115c2.slice/crio-1434282c860d25fb3748504a08e6ae32ac97e00742bd8215d9722863b807882b WatchSource:0}: Error finding container 1434282c860d25fb3748504a08e6ae32ac97e00742bd8215d9722863b807882b: Status 404 returned error can't find the container with id 1434282c860d25fb3748504a08e6ae32ac97e00742bd8215d9722863b807882b Sep 30 14:10:10 crc kubenswrapper[4840]: I0930 14:10:10.460687 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-tqjvl"] Sep 30 14:10:10 crc kubenswrapper[4840]: E0930 14:10:10.461630 4840 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-q6s8n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-589c58c6c-tqjvl_openstack-operators(3ed332eb-b604-4667-90f4-541d5a9115c2): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Sep 30 14:10:10 crc kubenswrapper[4840]: I0930 14:10:10.465442 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-t2mx4"] Sep 30 14:10:10 crc kubenswrapper[4840]: W0930 14:10:10.465788 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3d5da746_e5f3_4bed_9c77_37a78da8f1a6.slice/crio-5cd5d9e06639ee340ae83d61d0e3627680362b7243c417ace24fda0f6259b858 WatchSource:0}: Error finding container 5cd5d9e06639ee340ae83d61d0e3627680362b7243c417ace24fda0f6259b858: Status 404 returned error can't find the container with id 5cd5d9e06639ee340ae83d61d0e3627680362b7243c417ace24fda0f6259b858 Sep 30 14:10:10 crc kubenswrapper[4840]: I0930 14:10:10.469006 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64d7b59854-qsq84"] Sep 30 14:10:10 crc kubenswrapper[4840]: E0930 14:10:10.475957 4840 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:485df5c7813cdf4cf21f48ec48c8e3e4962fee6a1ae4c64f7af127d5ab346a10,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-k879m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-64d7b59854-qsq84_openstack-operators(3d5da746-e5f3-4bed-9c77-37a78da8f1a6): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Sep 30 14:10:10 crc kubenswrapper[4840]: I0930 14:10:10.585445 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rzkrh"] Sep 30 14:10:10 crc kubenswrapper[4840]: I0930 14:10:10.595874 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rzkrh" Sep 30 14:10:10 crc kubenswrapper[4840]: I0930 14:10:10.617972 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rzkrh"] Sep 30 14:10:10 crc kubenswrapper[4840]: I0930 14:10:10.640723 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ae5abcf-aeb1-4607-8285-cd6ea1d1f912-catalog-content\") pod \"redhat-operators-rzkrh\" (UID: \"0ae5abcf-aeb1-4607-8285-cd6ea1d1f912\") " pod="openshift-marketplace/redhat-operators-rzkrh" Sep 30 14:10:10 crc kubenswrapper[4840]: I0930 14:10:10.640769 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8s7b\" (UniqueName: \"kubernetes.io/projected/0ae5abcf-aeb1-4607-8285-cd6ea1d1f912-kube-api-access-q8s7b\") pod \"redhat-operators-rzkrh\" (UID: \"0ae5abcf-aeb1-4607-8285-cd6ea1d1f912\") " pod="openshift-marketplace/redhat-operators-rzkrh" Sep 30 14:10:10 crc kubenswrapper[4840]: I0930 14:10:10.640839 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ae5abcf-aeb1-4607-8285-cd6ea1d1f912-utilities\") pod \"redhat-operators-rzkrh\" (UID: \"0ae5abcf-aeb1-4607-8285-cd6ea1d1f912\") " pod="openshift-marketplace/redhat-operators-rzkrh" Sep 30 14:10:10 crc kubenswrapper[4840]: E0930 14:10:10.658974 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-qsq84" podUID="3d5da746-e5f3-4bed-9c77-37a78da8f1a6" Sep 30 14:10:10 crc kubenswrapper[4840]: E0930 14:10:10.659156 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-tqjvl" podUID="3ed332eb-b604-4667-90f4-541d5a9115c2" Sep 30 14:10:10 crc kubenswrapper[4840]: I0930 14:10:10.742756 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ae5abcf-aeb1-4607-8285-cd6ea1d1f912-catalog-content\") pod \"redhat-operators-rzkrh\" (UID: \"0ae5abcf-aeb1-4607-8285-cd6ea1d1f912\") " pod="openshift-marketplace/redhat-operators-rzkrh" Sep 30 14:10:10 crc kubenswrapper[4840]: I0930 14:10:10.742799 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8s7b\" (UniqueName: \"kubernetes.io/projected/0ae5abcf-aeb1-4607-8285-cd6ea1d1f912-kube-api-access-q8s7b\") pod \"redhat-operators-rzkrh\" (UID: \"0ae5abcf-aeb1-4607-8285-cd6ea1d1f912\") " pod="openshift-marketplace/redhat-operators-rzkrh" Sep 30 14:10:10 crc kubenswrapper[4840]: I0930 14:10:10.742875 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ae5abcf-aeb1-4607-8285-cd6ea1d1f912-utilities\") pod \"redhat-operators-rzkrh\" (UID: \"0ae5abcf-aeb1-4607-8285-cd6ea1d1f912\") " pod="openshift-marketplace/redhat-operators-rzkrh" Sep 30 14:10:10 crc kubenswrapper[4840]: I0930 14:10:10.743399 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ae5abcf-aeb1-4607-8285-cd6ea1d1f912-utilities\") pod \"redhat-operators-rzkrh\" (UID: \"0ae5abcf-aeb1-4607-8285-cd6ea1d1f912\") " pod="openshift-marketplace/redhat-operators-rzkrh" Sep 30 14:10:10 crc kubenswrapper[4840]: I0930 14:10:10.743593 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ae5abcf-aeb1-4607-8285-cd6ea1d1f912-catalog-content\") pod \"redhat-operators-rzkrh\" (UID: \"0ae5abcf-aeb1-4607-8285-cd6ea1d1f912\") " pod="openshift-marketplace/redhat-operators-rzkrh" Sep 30 14:10:10 crc kubenswrapper[4840]: I0930 14:10:10.769119 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8s7b\" (UniqueName: \"kubernetes.io/projected/0ae5abcf-aeb1-4607-8285-cd6ea1d1f912-kube-api-access-q8s7b\") pod \"redhat-operators-rzkrh\" (UID: \"0ae5abcf-aeb1-4607-8285-cd6ea1d1f912\") " pod="openshift-marketplace/redhat-operators-rzkrh" Sep 30 14:10:10 crc kubenswrapper[4840]: I0930 14:10:10.858973 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-78974455b9-97rs4"] Sep 30 14:10:10 crc kubenswrapper[4840]: I0930 14:10:10.864579 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-p6msl"] Sep 30 14:10:10 crc kubenswrapper[4840]: W0930 14:10:10.869303 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod807f2091_02fa_41e0_943a_b30f4202c476.slice/crio-e632a3515681aeddd95cb921cc049528ff5302e1ef098bda7016d8ed6ccb5770 WatchSource:0}: Error finding container e632a3515681aeddd95cb921cc049528ff5302e1ef098bda7016d8ed6ccb5770: Status 404 returned error can't find the container with id e632a3515681aeddd95cb921cc049528ff5302e1ef098bda7016d8ed6ccb5770 Sep 30 14:10:10 crc kubenswrapper[4840]: I0930 14:10:10.876498 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-76669f99c-rn2qq"] Sep 30 14:10:10 crc kubenswrapper[4840]: W0930 14:10:10.879740 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod62475ec3_b883_4d16_bb31_6ee393062da2.slice/crio-676a13dfd3916f99aa0ea8c72723fc8bb17dcf20c3b1b280bfd8f5983df123b5 WatchSource:0}: Error finding container 676a13dfd3916f99aa0ea8c72723fc8bb17dcf20c3b1b280bfd8f5983df123b5: Status 404 returned error can't find the container with id 676a13dfd3916f99aa0ea8c72723fc8bb17dcf20c3b1b280bfd8f5983df123b5 Sep 30 14:10:10 crc kubenswrapper[4840]: I0930 14:10:10.886363 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-smjg7"] Sep 30 14:10:10 crc kubenswrapper[4840]: E0930 14:10:10.892083 4840 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:225524223bf2a7f3a4ce95958fc9ca6fdab02745fb70374e8ff5bf1ddaceda4b,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-99j8v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-79d8469568-smjg7_openstack-operators(aa2cf241-ec41-4a94-b452-653b3439a073): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Sep 30 14:10:10 crc kubenswrapper[4840]: E0930 14:10:10.894439 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-smjg7" podUID="aa2cf241-ec41-4a94-b452-653b3439a073" Sep 30 14:10:10 crc kubenswrapper[4840]: I0930 14:10:10.901707 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-7d857cc749-xtqwd"] Sep 30 14:10:10 crc kubenswrapper[4840]: E0930 14:10:10.902664 4840 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.129.56.93:5001/openstack-k8s-operators/telemetry-operator:ac359d938872c47e1f3d7d8466b12f9d1f8a5236,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bj2jb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-78974455b9-97rs4_openstack-operators(47248483-6556-491b-b65f-01cad9f1d1fc): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Sep 30 14:10:10 crc kubenswrapper[4840]: W0930 14:10:10.905862 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3ba6da78_9606_4e1c_bc55_6cbe2e052464.slice/crio-5ceba9dc48f1c91cdb19c92f522f6e13ca886dffd0111002e09fd1b47b059695 WatchSource:0}: Error finding container 5ceba9dc48f1c91cdb19c92f522f6e13ca886dffd0111002e09fd1b47b059695: Status 404 returned error can't find the container with id 5ceba9dc48f1c91cdb19c92f522f6e13ca886dffd0111002e09fd1b47b059695 Sep 30 14:10:10 crc kubenswrapper[4840]: I0930 14:10:10.910041 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-f66b554c6-9npm5"] Sep 30 14:10:10 crc kubenswrapper[4840]: W0930 14:10:10.912511 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode1cd519e_b5c3_48f2_b426_d0f4e912683b.slice/crio-62e6c5371a288ed549d769a0487cd0d18987c691944cf90988b10ab5a7db2612 WatchSource:0}: Error finding container 62e6c5371a288ed549d769a0487cd0d18987c691944cf90988b10ab5a7db2612: Status 404 returned error can't find the container with id 62e6c5371a288ed549d769a0487cd0d18987c691944cf90988b10ab5a7db2612 Sep 30 14:10:10 crc kubenswrapper[4840]: E0930 14:10:10.921817 4840 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:a303e460aec09217f90043b8ff19c01061af003b614833b33a593df9c00ddf80,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-v98vj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-f66b554c6-9npm5_openstack-operators(3ba6da78-9606-4e1c-bc55-6cbe2e052464): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Sep 30 14:10:10 crc kubenswrapper[4840]: E0930 14:10:10.924108 4840 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:7169dfadf5f5589f14ca52700d2eba991c2a0c7733f6a1ea795752d993d7f61b,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fv6j5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-76669f99c-rn2qq_openstack-operators(e1cd519e-b5c3-48f2-b426-d0f4e912683b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Sep 30 14:10:10 crc kubenswrapper[4840]: I0930 14:10:10.931760 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rzkrh" Sep 30 14:10:10 crc kubenswrapper[4840]: I0930 14:10:10.939251 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-568796b78d-bvbdn"] Sep 30 14:10:10 crc kubenswrapper[4840]: W0930 14:10:10.982699 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode61dcaab_7705_4857_bdd9_adeb07d9e6f2.slice/crio-f47446122d4d729c025df0f72003f514343856c49cabe47f5fa083c3455d95d9 WatchSource:0}: Error finding container f47446122d4d729c025df0f72003f514343856c49cabe47f5fa083c3455d95d9: Status 404 returned error can't find the container with id f47446122d4d729c025df0f72003f514343856c49cabe47f5fa083c3455d95d9 Sep 30 14:10:11 crc kubenswrapper[4840]: E0930 14:10:11.132811 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-f66b554c6-9npm5" podUID="3ba6da78-9606-4e1c-bc55-6cbe2e052464" Sep 30 14:10:11 crc kubenswrapper[4840]: E0930 14:10:11.146850 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-78974455b9-97rs4" podUID="47248483-6556-491b-b65f-01cad9f1d1fc" Sep 30 14:10:11 crc kubenswrapper[4840]: I0930 14:10:11.217788 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-4mpb8" event={"ID":"2b8703c9-f691-4c3a-9941-8bcd36aaf948","Type":"ContainerStarted","Data":"836a86f599e3ba47c8102a6fed2558ab59dd04e175517f182396d665acfdbd82"} Sep 30 14:10:11 crc kubenswrapper[4840]: I0930 14:10:11.222266 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-qsq84" event={"ID":"3d5da746-e5f3-4bed-9c77-37a78da8f1a6","Type":"ContainerStarted","Data":"7c6339dd26cb9608548904e9d43e2b70da67775365c912e80bb359165af9f197"} Sep 30 14:10:11 crc kubenswrapper[4840]: I0930 14:10:11.222356 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-qsq84" event={"ID":"3d5da746-e5f3-4bed-9c77-37a78da8f1a6","Type":"ContainerStarted","Data":"5cd5d9e06639ee340ae83d61d0e3627680362b7243c417ace24fda0f6259b858"} Sep 30 14:10:11 crc kubenswrapper[4840]: E0930 14:10:11.225068 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-rn2qq" podUID="e1cd519e-b5c3-48f2-b426-d0f4e912683b" Sep 30 14:10:11 crc kubenswrapper[4840]: E0930 14:10:11.231492 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:485df5c7813cdf4cf21f48ec48c8e3e4962fee6a1ae4c64f7af127d5ab346a10\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-qsq84" podUID="3d5da746-e5f3-4bed-9c77-37a78da8f1a6" Sep 30 14:10:11 crc kubenswrapper[4840]: I0930 14:10:11.232671 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-xtqwd" event={"ID":"807f2091-02fa-41e0-943a-b30f4202c476","Type":"ContainerStarted","Data":"e632a3515681aeddd95cb921cc049528ff5302e1ef098bda7016d8ed6ccb5770"} Sep 30 14:10:11 crc kubenswrapper[4840]: I0930 14:10:11.233513 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-p6msl" event={"ID":"62475ec3-b883-4d16-bb31-6ee393062da2","Type":"ContainerStarted","Data":"676a13dfd3916f99aa0ea8c72723fc8bb17dcf20c3b1b280bfd8f5983df123b5"} Sep 30 14:10:11 crc kubenswrapper[4840]: I0930 14:10:11.240744 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-rn2qq" event={"ID":"e1cd519e-b5c3-48f2-b426-d0f4e912683b","Type":"ContainerStarted","Data":"62e6c5371a288ed549d769a0487cd0d18987c691944cf90988b10ab5a7db2612"} Sep 30 14:10:11 crc kubenswrapper[4840]: E0930 14:10:11.244136 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:7169dfadf5f5589f14ca52700d2eba991c2a0c7733f6a1ea795752d993d7f61b\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-rn2qq" podUID="e1cd519e-b5c3-48f2-b426-d0f4e912683b" Sep 30 14:10:11 crc kubenswrapper[4840]: I0930 14:10:11.254103 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-d2hgd" event={"ID":"eb939f27-ce94-408d-b673-4c25f96dd059","Type":"ContainerStarted","Data":"e13626b5ac120cd1401475f7ad14ef8d06d6acd1d8d37d938a2c224b4e4a514b"} Sep 30 14:10:11 crc kubenswrapper[4840]: I0930 14:10:11.261195 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-smjg7" event={"ID":"aa2cf241-ec41-4a94-b452-653b3439a073","Type":"ContainerStarted","Data":"b2f44ea6fa224b83956ff0b6b77aabe4709329945298bc1b88ae8319e0dae495"} Sep 30 14:10:11 crc kubenswrapper[4840]: E0930 14:10:11.265518 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:225524223bf2a7f3a4ce95958fc9ca6fdab02745fb70374e8ff5bf1ddaceda4b\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-smjg7" podUID="aa2cf241-ec41-4a94-b452-653b3439a073" Sep 30 14:10:11 crc kubenswrapper[4840]: I0930 14:10:11.269792 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-sq5hd" event={"ID":"64c5ad91-f113-46e3-ac87-36521d32f4f0","Type":"ContainerStarted","Data":"a5b1127a38bc81885a8e4c44692546a3c29ed1373b333dce46d1e77f9e131f2e"} Sep 30 14:10:11 crc kubenswrapper[4840]: I0930 14:10:11.272832 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-568796b78d-bvbdn" event={"ID":"e61dcaab-7705-4857-bdd9-adeb07d9e6f2","Type":"ContainerStarted","Data":"f47446122d4d729c025df0f72003f514343856c49cabe47f5fa083c3455d95d9"} Sep 30 14:10:11 crc kubenswrapper[4840]: I0930 14:10:11.282161 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-78974455b9-97rs4" event={"ID":"47248483-6556-491b-b65f-01cad9f1d1fc","Type":"ContainerStarted","Data":"d4c39963765ad9f24b7b169b86ba6d2cabae1605bd9ee8c52d88ea812d8c4ff8"} Sep 30 14:10:11 crc kubenswrapper[4840]: I0930 14:10:11.282199 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-78974455b9-97rs4" event={"ID":"47248483-6556-491b-b65f-01cad9f1d1fc","Type":"ContainerStarted","Data":"0b68bc183bd9ff9ebdbd989d67e3d9ffd8ed13df962c52a6b4ae610229e38269"} Sep 30 14:10:11 crc kubenswrapper[4840]: I0930 14:10:11.289658 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-f66b554c6-9npm5" event={"ID":"3ba6da78-9606-4e1c-bc55-6cbe2e052464","Type":"ContainerStarted","Data":"680f7f782ad5b2e7c589c496b24c1fb248f809f9e5d3658735d0ffa7fd9c81b0"} Sep 30 14:10:11 crc kubenswrapper[4840]: I0930 14:10:11.289909 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-f66b554c6-9npm5" event={"ID":"3ba6da78-9606-4e1c-bc55-6cbe2e052464","Type":"ContainerStarted","Data":"5ceba9dc48f1c91cdb19c92f522f6e13ca886dffd0111002e09fd1b47b059695"} Sep 30 14:10:11 crc kubenswrapper[4840]: I0930 14:10:11.293026 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-9t25m" event={"ID":"dc75c0e8-691e-4108-b9c7-e4233d0baeb6","Type":"ContainerStarted","Data":"8e45575138ac2e33ec4b2a54d01152d1c484b24152856490a41f89a269f5bd31"} Sep 30 14:10:11 crc kubenswrapper[4840]: I0930 14:10:11.327774 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-rjpbm" event={"ID":"b927784b-0b44-4bf1-9751-3aafcc24367b","Type":"ContainerStarted","Data":"a8322b858cb54273769d8e292136843422c6fdad4f76d245749f98e6d596e62d"} Sep 30 14:10:11 crc kubenswrapper[4840]: I0930 14:10:11.345189 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-jwlj9" event={"ID":"5daf5662-8848-4fd9-a6af-e091fe303d89","Type":"ContainerStarted","Data":"6aee594b0466d52683ccea872b61e1c661883543fcef6761f06f3b7c556a39e9"} Sep 30 14:10:11 crc kubenswrapper[4840]: I0930 14:10:11.350199 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-t2mx4" event={"ID":"a76ab4ec-b180-473f-a597-641725526b58","Type":"ContainerStarted","Data":"d90d64e6cd7098806ffbcdc8324c9b26488804248bde891b7f5122f4c231a204"} Sep 30 14:10:11 crc kubenswrapper[4840]: I0930 14:10:11.362643 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-tqjvl" event={"ID":"3ed332eb-b604-4667-90f4-541d5a9115c2","Type":"ContainerStarted","Data":"3dc4d38e9b86c701b53d60a7124dded8fea97c1e52d5b0490ee874d24ffbe59d"} Sep 30 14:10:11 crc kubenswrapper[4840]: I0930 14:10:11.362678 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-tqjvl" event={"ID":"3ed332eb-b604-4667-90f4-541d5a9115c2","Type":"ContainerStarted","Data":"1434282c860d25fb3748504a08e6ae32ac97e00742bd8215d9722863b807882b"} Sep 30 14:10:11 crc kubenswrapper[4840]: I0930 14:10:11.365023 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-fg7c7" event={"ID":"494a382c-6fec-4711-8a90-7f691c847e86","Type":"ContainerStarted","Data":"75ab28e60227233fc893574661577ecc831454d1acc8671e084ba69043a9dba9"} Sep 30 14:10:11 crc kubenswrapper[4840]: E0930 14:10:11.376191 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2\\\"\"" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-tqjvl" podUID="3ed332eb-b604-4667-90f4-541d5a9115c2" Sep 30 14:10:11 crc kubenswrapper[4840]: E0930 14:10:11.376284 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.93:5001/openstack-k8s-operators/telemetry-operator:ac359d938872c47e1f3d7d8466b12f9d1f8a5236\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-78974455b9-97rs4" podUID="47248483-6556-491b-b65f-01cad9f1d1fc" Sep 30 14:10:11 crc kubenswrapper[4840]: E0930 14:10:11.376348 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:a303e460aec09217f90043b8ff19c01061af003b614833b33a593df9c00ddf80\\\"\"" pod="openstack-operators/test-operator-controller-manager-f66b554c6-9npm5" podUID="3ba6da78-9606-4e1c-bc55-6cbe2e052464" Sep 30 14:10:11 crc kubenswrapper[4840]: I0930 14:10:11.383872 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-d9jgv" event={"ID":"63d94c36-0a55-472c-bbe8-27195105584e","Type":"ContainerStarted","Data":"3a88bb51e4fa1193f4fca62ffdafd109c5584f344c59390a3c534d8acd431842"} Sep 30 14:10:11 crc kubenswrapper[4840]: I0930 14:10:11.389306 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-x4mx4" event={"ID":"3f4d112b-e6b8-455a-9f66-e8ee1bb81ac9","Type":"ContainerStarted","Data":"c12958c12cca5853c113820864815f780898a212b80fe617ce1149d962cebd9a"} Sep 30 14:10:11 crc kubenswrapper[4840]: I0930 14:10:11.392186 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-xzmg8" event={"ID":"385365f0-9e2c-4c89-b95b-164a473ff0c3","Type":"ContainerStarted","Data":"0ce69348b92f7a0a400343740e4f7d383be05503ce9e50c31fc12b28d4a494cb"} Sep 30 14:10:11 crc kubenswrapper[4840]: I0930 14:10:11.557894 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rzkrh"] Sep 30 14:10:12 crc kubenswrapper[4840]: I0930 14:10:12.445800 4840 generic.go:334] "Generic (PLEG): container finished" podID="0ae5abcf-aeb1-4607-8285-cd6ea1d1f912" containerID="68ea64bdefdbc9989c3d91732a5e72024e8d80596b7052e8471bce204a775601" exitCode=0 Sep 30 14:10:12 crc kubenswrapper[4840]: I0930 14:10:12.445975 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rzkrh" event={"ID":"0ae5abcf-aeb1-4607-8285-cd6ea1d1f912","Type":"ContainerDied","Data":"68ea64bdefdbc9989c3d91732a5e72024e8d80596b7052e8471bce204a775601"} Sep 30 14:10:12 crc kubenswrapper[4840]: I0930 14:10:12.446169 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rzkrh" event={"ID":"0ae5abcf-aeb1-4607-8285-cd6ea1d1f912","Type":"ContainerStarted","Data":"5fff5aa57c824f4360c7d206b723e4d6396d9687d3538fb33cf1c0d9ab50aa15"} Sep 30 14:10:12 crc kubenswrapper[4840]: I0930 14:10:12.454896 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-568796b78d-bvbdn" event={"ID":"e61dcaab-7705-4857-bdd9-adeb07d9e6f2","Type":"ContainerStarted","Data":"0887e6a78847ab6f2495b9ba9c0b3a0621684ec174687bc683e4a3689de4e60f"} Sep 30 14:10:12 crc kubenswrapper[4840]: I0930 14:10:12.454949 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-568796b78d-bvbdn" event={"ID":"e61dcaab-7705-4857-bdd9-adeb07d9e6f2","Type":"ContainerStarted","Data":"8af3d054057ef8049d5c82268bec1e5e13270461d328885b0d86c8216dd58900"} Sep 30 14:10:12 crc kubenswrapper[4840]: I0930 14:10:12.455804 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-568796b78d-bvbdn" Sep 30 14:10:12 crc kubenswrapper[4840]: I0930 14:10:12.465224 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-rn2qq" event={"ID":"e1cd519e-b5c3-48f2-b426-d0f4e912683b","Type":"ContainerStarted","Data":"485572b84d949ad87c9fcf0ab3018ac01062865d781856b65cbec45b3ff67ec2"} Sep 30 14:10:12 crc kubenswrapper[4840]: E0930 14:10:12.512671 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:225524223bf2a7f3a4ce95958fc9ca6fdab02745fb70374e8ff5bf1ddaceda4b\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-smjg7" podUID="aa2cf241-ec41-4a94-b452-653b3439a073" Sep 30 14:10:12 crc kubenswrapper[4840]: E0930 14:10:12.512848 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:a303e460aec09217f90043b8ff19c01061af003b614833b33a593df9c00ddf80\\\"\"" pod="openstack-operators/test-operator-controller-manager-f66b554c6-9npm5" podUID="3ba6da78-9606-4e1c-bc55-6cbe2e052464" Sep 30 14:10:12 crc kubenswrapper[4840]: E0930 14:10:12.513307 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.93:5001/openstack-k8s-operators/telemetry-operator:ac359d938872c47e1f3d7d8466b12f9d1f8a5236\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-78974455b9-97rs4" podUID="47248483-6556-491b-b65f-01cad9f1d1fc" Sep 30 14:10:12 crc kubenswrapper[4840]: E0930 14:10:12.513442 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:7169dfadf5f5589f14ca52700d2eba991c2a0c7733f6a1ea795752d993d7f61b\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-rn2qq" podUID="e1cd519e-b5c3-48f2-b426-d0f4e912683b" Sep 30 14:10:12 crc kubenswrapper[4840]: E0930 14:10:12.513597 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:485df5c7813cdf4cf21f48ec48c8e3e4962fee6a1ae4c64f7af127d5ab346a10\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-qsq84" podUID="3d5da746-e5f3-4bed-9c77-37a78da8f1a6" Sep 30 14:10:12 crc kubenswrapper[4840]: E0930 14:10:12.513760 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2\\\"\"" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-tqjvl" podUID="3ed332eb-b604-4667-90f4-541d5a9115c2" Sep 30 14:10:12 crc kubenswrapper[4840]: I0930 14:10:12.551421 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-568796b78d-bvbdn" podStartSLOduration=3.551407573 podStartE2EDuration="3.551407573s" podCreationTimestamp="2025-09-30 14:10:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:10:12.548239342 +0000 UTC m=+841.177325765" watchObservedRunningTime="2025-09-30 14:10:12.551407573 +0000 UTC m=+841.180493996" Sep 30 14:10:13 crc kubenswrapper[4840]: E0930 14:10:13.472217 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:7169dfadf5f5589f14ca52700d2eba991c2a0c7733f6a1ea795752d993d7f61b\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-rn2qq" podUID="e1cd519e-b5c3-48f2-b426-d0f4e912683b" Sep 30 14:10:20 crc kubenswrapper[4840]: I0930 14:10:20.462460 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-568796b78d-bvbdn" Sep 30 14:10:21 crc kubenswrapper[4840]: I0930 14:10:21.871599 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:10:21 crc kubenswrapper[4840]: I0930 14:10:21.871929 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:10:21 crc kubenswrapper[4840]: I0930 14:10:21.871967 4840 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-747gk" Sep 30 14:10:21 crc kubenswrapper[4840]: I0930 14:10:21.872481 4840 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"999ab6436a76ee1df695f7a3673950dda45d0789af244839074f28ed47e2534b"} pod="openshift-machine-config-operator/machine-config-daemon-747gk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 14:10:21 crc kubenswrapper[4840]: I0930 14:10:21.872539 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" containerID="cri-o://999ab6436a76ee1df695f7a3673950dda45d0789af244839074f28ed47e2534b" gracePeriod=600 Sep 30 14:10:22 crc kubenswrapper[4840]: E0930 14:10:22.019275 4840 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10e8b890_7f20_4a36_8e03_898620cf599a.slice/crio-999ab6436a76ee1df695f7a3673950dda45d0789af244839074f28ed47e2534b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10e8b890_7f20_4a36_8e03_898620cf599a.slice/crio-conmon-999ab6436a76ee1df695f7a3673950dda45d0789af244839074f28ed47e2534b.scope\": RecentStats: unable to find data in memory cache]" Sep 30 14:10:22 crc kubenswrapper[4840]: I0930 14:10:22.529056 4840 generic.go:334] "Generic (PLEG): container finished" podID="10e8b890-7f20-4a36-8e03-898620cf599a" containerID="999ab6436a76ee1df695f7a3673950dda45d0789af244839074f28ed47e2534b" exitCode=0 Sep 30 14:10:22 crc kubenswrapper[4840]: I0930 14:10:22.529104 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" event={"ID":"10e8b890-7f20-4a36-8e03-898620cf599a","Type":"ContainerDied","Data":"999ab6436a76ee1df695f7a3673950dda45d0789af244839074f28ed47e2534b"} Sep 30 14:10:22 crc kubenswrapper[4840]: I0930 14:10:22.529154 4840 scope.go:117] "RemoveContainer" containerID="d12912f690026de70f88356c602b1439b6380a977591bd20b79ac057d9633343" Sep 30 14:10:28 crc kubenswrapper[4840]: I0930 14:10:28.246033 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-47pd9"] Sep 30 14:10:28 crc kubenswrapper[4840]: I0930 14:10:28.248341 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-47pd9" Sep 30 14:10:28 crc kubenswrapper[4840]: I0930 14:10:28.257221 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-47pd9"] Sep 30 14:10:28 crc kubenswrapper[4840]: I0930 14:10:28.349138 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3d4cfc5-2801-47a0-9c44-1943efdaa0ae-catalog-content\") pod \"redhat-marketplace-47pd9\" (UID: \"d3d4cfc5-2801-47a0-9c44-1943efdaa0ae\") " pod="openshift-marketplace/redhat-marketplace-47pd9" Sep 30 14:10:28 crc kubenswrapper[4840]: I0930 14:10:28.349212 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3d4cfc5-2801-47a0-9c44-1943efdaa0ae-utilities\") pod \"redhat-marketplace-47pd9\" (UID: \"d3d4cfc5-2801-47a0-9c44-1943efdaa0ae\") " pod="openshift-marketplace/redhat-marketplace-47pd9" Sep 30 14:10:28 crc kubenswrapper[4840]: I0930 14:10:28.349231 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nntbs\" (UniqueName: \"kubernetes.io/projected/d3d4cfc5-2801-47a0-9c44-1943efdaa0ae-kube-api-access-nntbs\") pod \"redhat-marketplace-47pd9\" (UID: \"d3d4cfc5-2801-47a0-9c44-1943efdaa0ae\") " pod="openshift-marketplace/redhat-marketplace-47pd9" Sep 30 14:10:28 crc kubenswrapper[4840]: I0930 14:10:28.450872 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3d4cfc5-2801-47a0-9c44-1943efdaa0ae-catalog-content\") pod \"redhat-marketplace-47pd9\" (UID: \"d3d4cfc5-2801-47a0-9c44-1943efdaa0ae\") " pod="openshift-marketplace/redhat-marketplace-47pd9" Sep 30 14:10:28 crc kubenswrapper[4840]: I0930 14:10:28.451196 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3d4cfc5-2801-47a0-9c44-1943efdaa0ae-utilities\") pod \"redhat-marketplace-47pd9\" (UID: \"d3d4cfc5-2801-47a0-9c44-1943efdaa0ae\") " pod="openshift-marketplace/redhat-marketplace-47pd9" Sep 30 14:10:28 crc kubenswrapper[4840]: I0930 14:10:28.451282 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nntbs\" (UniqueName: \"kubernetes.io/projected/d3d4cfc5-2801-47a0-9c44-1943efdaa0ae-kube-api-access-nntbs\") pod \"redhat-marketplace-47pd9\" (UID: \"d3d4cfc5-2801-47a0-9c44-1943efdaa0ae\") " pod="openshift-marketplace/redhat-marketplace-47pd9" Sep 30 14:10:28 crc kubenswrapper[4840]: I0930 14:10:28.451640 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3d4cfc5-2801-47a0-9c44-1943efdaa0ae-utilities\") pod \"redhat-marketplace-47pd9\" (UID: \"d3d4cfc5-2801-47a0-9c44-1943efdaa0ae\") " pod="openshift-marketplace/redhat-marketplace-47pd9" Sep 30 14:10:28 crc kubenswrapper[4840]: I0930 14:10:28.451962 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3d4cfc5-2801-47a0-9c44-1943efdaa0ae-catalog-content\") pod \"redhat-marketplace-47pd9\" (UID: \"d3d4cfc5-2801-47a0-9c44-1943efdaa0ae\") " pod="openshift-marketplace/redhat-marketplace-47pd9" Sep 30 14:10:28 crc kubenswrapper[4840]: I0930 14:10:28.472764 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nntbs\" (UniqueName: \"kubernetes.io/projected/d3d4cfc5-2801-47a0-9c44-1943efdaa0ae-kube-api-access-nntbs\") pod \"redhat-marketplace-47pd9\" (UID: \"d3d4cfc5-2801-47a0-9c44-1943efdaa0ae\") " pod="openshift-marketplace/redhat-marketplace-47pd9" Sep 30 14:10:28 crc kubenswrapper[4840]: I0930 14:10:28.564866 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-47pd9" Sep 30 14:10:32 crc kubenswrapper[4840]: I0930 14:10:32.613068 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-9t25m" event={"ID":"dc75c0e8-691e-4108-b9c7-e4233d0baeb6","Type":"ContainerStarted","Data":"4667da904036e49a79e8249e614e5e03b929931ebb8a162eb54e14e81e8ec992"} Sep 30 14:10:32 crc kubenswrapper[4840]: I0930 14:10:32.617623 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-6nn29" event={"ID":"395025e2-9d12-4172-bb73-b4fb67e0111c","Type":"ContainerStarted","Data":"97cad3e907396acd8e1e4b3ac5d79c65e81584eadc7b4350cec3622f3430caed"} Sep 30 14:10:32 crc kubenswrapper[4840]: I0930 14:10:32.622845 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-sq5hd" event={"ID":"64c5ad91-f113-46e3-ac87-36521d32f4f0","Type":"ContainerStarted","Data":"82854d19ed04d1bae2042f47876f03627068b60e292b9deadd61d8472d7866af"} Sep 30 14:10:32 crc kubenswrapper[4840]: I0930 14:10:32.628318 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" event={"ID":"10e8b890-7f20-4a36-8e03-898620cf599a","Type":"ContainerStarted","Data":"6060b9705d9c27f06e58cff4fd0a3ee141a9c798fa228092898474533d2401f1"} Sep 30 14:10:32 crc kubenswrapper[4840]: I0930 14:10:32.638579 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-fg7c7" event={"ID":"494a382c-6fec-4711-8a90-7f691c847e86","Type":"ContainerStarted","Data":"5219f6138d6599d4281815e8362f693e1d7075c4f24b77cb1f00915df650eaf8"} Sep 30 14:10:32 crc kubenswrapper[4840]: I0930 14:10:32.689067 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-47pd9"] Sep 30 14:10:32 crc kubenswrapper[4840]: W0930 14:10:32.715071 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd3d4cfc5_2801_47a0_9c44_1943efdaa0ae.slice/crio-7b73a3aa7e7eca3fb024fd1da37e42577e2cd4c262c0a8811c28068bb597acce WatchSource:0}: Error finding container 7b73a3aa7e7eca3fb024fd1da37e42577e2cd4c262c0a8811c28068bb597acce: Status 404 returned error can't find the container with id 7b73a3aa7e7eca3fb024fd1da37e42577e2cd4c262c0a8811c28068bb597acce Sep 30 14:10:33 crc kubenswrapper[4840]: I0930 14:10:33.672247 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-jwlj9" event={"ID":"5daf5662-8848-4fd9-a6af-e091fe303d89","Type":"ContainerStarted","Data":"7312c013d87c1c777c7f71118b441074e89d269c21d441fdaa5f8a3c0443f7b7"} Sep 30 14:10:33 crc kubenswrapper[4840]: I0930 14:10:33.690100 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rzkrh" event={"ID":"0ae5abcf-aeb1-4607-8285-cd6ea1d1f912","Type":"ContainerStarted","Data":"68894737287ff74090f704c5e3d6492c424a762a23ec926b3cd2f0009eaed1a2"} Sep 30 14:10:33 crc kubenswrapper[4840]: I0930 14:10:33.709666 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-4mpb8" event={"ID":"2b8703c9-f691-4c3a-9941-8bcd36aaf948","Type":"ContainerStarted","Data":"df43a49c29cd82231fe1d616af2238948e3cd8990d4c92d78d3e97e16d76a341"} Sep 30 14:10:33 crc kubenswrapper[4840]: I0930 14:10:33.715911 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-d9jgv" event={"ID":"63d94c36-0a55-472c-bbe8-27195105584e","Type":"ContainerStarted","Data":"f29a7811c7e7e22cfac44b33b7e53b345f1d3b9e18037ce0440c1fe497ef1991"} Sep 30 14:10:33 crc kubenswrapper[4840]: I0930 14:10:33.718634 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-xzmg8" event={"ID":"385365f0-9e2c-4c89-b95b-164a473ff0c3","Type":"ContainerStarted","Data":"5a04551d66b672ec2076d3c11bc82219ffa78c99657ce3a51accf04c53fd8cfa"} Sep 30 14:10:33 crc kubenswrapper[4840]: I0930 14:10:33.736832 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-smg6p" event={"ID":"3ffb0dd0-3dd8-4f37-8ead-4add746a1fc6","Type":"ContainerStarted","Data":"90b557b61da854a1c054907d0232d33089ddd409cd3bc5fb4ec81882d75fe831"} Sep 30 14:10:33 crc kubenswrapper[4840]: I0930 14:10:33.755332 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-rjpbm" event={"ID":"b927784b-0b44-4bf1-9751-3aafcc24367b","Type":"ContainerStarted","Data":"9b587e9808dbe2d13d5e950e005a33712f60edec2ca3424ada1ebc2dfc017455"} Sep 30 14:10:33 crc kubenswrapper[4840]: I0930 14:10:33.758376 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-d2hgd" event={"ID":"eb939f27-ce94-408d-b673-4c25f96dd059","Type":"ContainerStarted","Data":"03dd358246335b23d81f454363d0a481d31acf37691a140eb80adb5e8108988e"} Sep 30 14:10:33 crc kubenswrapper[4840]: I0930 14:10:33.759521 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-t2mx4" event={"ID":"a76ab4ec-b180-473f-a597-641725526b58","Type":"ContainerStarted","Data":"c57829410c303c40a9bf107e4417f5ed908d904e662ae71b6c5696dd23765464"} Sep 30 14:10:33 crc kubenswrapper[4840]: I0930 14:10:33.760464 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-p6msl" event={"ID":"62475ec3-b883-4d16-bb31-6ee393062da2","Type":"ContainerStarted","Data":"8fa786e57f858f86946298c94715ae4025ec01aa86ab847dea3391fc5e7c10db"} Sep 30 14:10:33 crc kubenswrapper[4840]: I0930 14:10:33.770776 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-rnsbn" event={"ID":"0f4e505b-4892-4c92-9ec4-0bdeb5d4c873","Type":"ContainerStarted","Data":"a2bcdc784ff7a3348d79297227f01e849cff29d326f9dfe3b330aa4acca490b5"} Sep 30 14:10:33 crc kubenswrapper[4840]: I0930 14:10:33.799136 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-x4mx4" event={"ID":"3f4d112b-e6b8-455a-9f66-e8ee1bb81ac9","Type":"ContainerStarted","Data":"dbb1fdc6d55058488334af9e39e208f345d58dbdf6a11a35dffe6b6deec6a1c3"} Sep 30 14:10:33 crc kubenswrapper[4840]: I0930 14:10:33.802344 4840 generic.go:334] "Generic (PLEG): container finished" podID="d3d4cfc5-2801-47a0-9c44-1943efdaa0ae" containerID="b1ab7febc7328622d487a292b00d0c98e61f871e6133ff2d908ca619c3ad25d8" exitCode=0 Sep 30 14:10:33 crc kubenswrapper[4840]: I0930 14:10:33.802408 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-47pd9" event={"ID":"d3d4cfc5-2801-47a0-9c44-1943efdaa0ae","Type":"ContainerDied","Data":"b1ab7febc7328622d487a292b00d0c98e61f871e6133ff2d908ca619c3ad25d8"} Sep 30 14:10:33 crc kubenswrapper[4840]: I0930 14:10:33.802430 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-47pd9" event={"ID":"d3d4cfc5-2801-47a0-9c44-1943efdaa0ae","Type":"ContainerStarted","Data":"7b73a3aa7e7eca3fb024fd1da37e42577e2cd4c262c0a8811c28068bb597acce"} Sep 30 14:10:33 crc kubenswrapper[4840]: I0930 14:10:33.806797 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-6nn29" event={"ID":"395025e2-9d12-4172-bb73-b4fb67e0111c","Type":"ContainerStarted","Data":"832331d6fdc3c30d6a147ddfbf7377e6b81d6756d4801e4ac53d9b9e2befc219"} Sep 30 14:10:33 crc kubenswrapper[4840]: I0930 14:10:33.807409 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-6nn29" Sep 30 14:10:33 crc kubenswrapper[4840]: I0930 14:10:33.818306 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-xtqwd" event={"ID":"807f2091-02fa-41e0-943a-b30f4202c476","Type":"ContainerStarted","Data":"15763bdc626a27db672f60519b9578fdf9b0a38de6ba205b52b315eb13e1f015"} Sep 30 14:10:33 crc kubenswrapper[4840]: I0930 14:10:33.866807 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-6nn29" podStartSLOduration=11.343080378 podStartE2EDuration="25.866766183s" podCreationTimestamp="2025-09-30 14:10:08 +0000 UTC" firstStartedPulling="2025-09-30 14:10:09.974092068 +0000 UTC m=+838.603178491" lastFinishedPulling="2025-09-30 14:10:24.497777863 +0000 UTC m=+853.126864296" observedRunningTime="2025-09-30 14:10:33.865713845 +0000 UTC m=+862.494800268" watchObservedRunningTime="2025-09-30 14:10:33.866766183 +0000 UTC m=+862.495852596" Sep 30 14:10:34 crc kubenswrapper[4840]: I0930 14:10:34.826412 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-rnsbn" event={"ID":"0f4e505b-4892-4c92-9ec4-0bdeb5d4c873","Type":"ContainerStarted","Data":"bed5b484d9420dd1fb97d5522cf3b71e8a7ed8613da5ec53645cbdb37b70edce"} Sep 30 14:10:34 crc kubenswrapper[4840]: I0930 14:10:34.827061 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-rnsbn" Sep 30 14:10:34 crc kubenswrapper[4840]: I0930 14:10:34.833505 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-x4mx4" event={"ID":"3f4d112b-e6b8-455a-9f66-e8ee1bb81ac9","Type":"ContainerStarted","Data":"421d8e4ba61ad9902a0de5dfadca7c19a32ab1203e9034f213f6a809cb5da345"} Sep 30 14:10:34 crc kubenswrapper[4840]: I0930 14:10:34.833667 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-x4mx4" Sep 30 14:10:34 crc kubenswrapper[4840]: I0930 14:10:34.838982 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-rjpbm" event={"ID":"b927784b-0b44-4bf1-9751-3aafcc24367b","Type":"ContainerStarted","Data":"9e84d320b07fc1d96a411caca6167fe7a3107be27e9e5fa1441359fff9ec643c"} Sep 30 14:10:34 crc kubenswrapper[4840]: I0930 14:10:34.841253 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-rjpbm" Sep 30 14:10:34 crc kubenswrapper[4840]: I0930 14:10:34.843739 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-rnsbn" podStartSLOduration=12.002395777 podStartE2EDuration="26.843722845s" podCreationTimestamp="2025-09-30 14:10:08 +0000 UTC" firstStartedPulling="2025-09-30 14:10:09.66585542 +0000 UTC m=+838.294941843" lastFinishedPulling="2025-09-30 14:10:24.507182488 +0000 UTC m=+853.136268911" observedRunningTime="2025-09-30 14:10:34.841020755 +0000 UTC m=+863.470107178" watchObservedRunningTime="2025-09-30 14:10:34.843722845 +0000 UTC m=+863.472809268" Sep 30 14:10:34 crc kubenswrapper[4840]: I0930 14:10:34.845149 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-xtqwd" event={"ID":"807f2091-02fa-41e0-943a-b30f4202c476","Type":"ContainerStarted","Data":"fa463818f19cac27b900f2dea298ec4c94bd8b72f075bfd7ea94ba044554b360"} Sep 30 14:10:34 crc kubenswrapper[4840]: I0930 14:10:34.845407 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-xtqwd" Sep 30 14:10:34 crc kubenswrapper[4840]: I0930 14:10:34.848375 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-9t25m" event={"ID":"dc75c0e8-691e-4108-b9c7-e4233d0baeb6","Type":"ContainerStarted","Data":"2d4f05d46be9c89e51e1371df824660273c3284f2cf50a801e68a481a360b868"} Sep 30 14:10:34 crc kubenswrapper[4840]: I0930 14:10:34.848591 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-9t25m" Sep 30 14:10:34 crc kubenswrapper[4840]: I0930 14:10:34.852692 4840 generic.go:334] "Generic (PLEG): container finished" podID="0ae5abcf-aeb1-4607-8285-cd6ea1d1f912" containerID="68894737287ff74090f704c5e3d6492c424a762a23ec926b3cd2f0009eaed1a2" exitCode=0 Sep 30 14:10:34 crc kubenswrapper[4840]: I0930 14:10:34.852738 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rzkrh" event={"ID":"0ae5abcf-aeb1-4607-8285-cd6ea1d1f912","Type":"ContainerDied","Data":"68894737287ff74090f704c5e3d6492c424a762a23ec926b3cd2f0009eaed1a2"} Sep 30 14:10:34 crc kubenswrapper[4840]: I0930 14:10:34.860431 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-rjpbm" podStartSLOduration=4.867020999 podStartE2EDuration="26.86041302s" podCreationTimestamp="2025-09-30 14:10:08 +0000 UTC" firstStartedPulling="2025-09-30 14:10:10.226434608 +0000 UTC m=+838.855521031" lastFinishedPulling="2025-09-30 14:10:32.219826629 +0000 UTC m=+860.848913052" observedRunningTime="2025-09-30 14:10:34.853624293 +0000 UTC m=+863.482710716" watchObservedRunningTime="2025-09-30 14:10:34.86041302 +0000 UTC m=+863.489499443" Sep 30 14:10:34 crc kubenswrapper[4840]: I0930 14:10:34.865110 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-d2hgd" event={"ID":"eb939f27-ce94-408d-b673-4c25f96dd059","Type":"ContainerStarted","Data":"cdad221d0651b74cc660003f68c04b3cd557b635b4a7ea8a3615a9e9d396ce82"} Sep 30 14:10:34 crc kubenswrapper[4840]: I0930 14:10:34.865380 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-d2hgd" Sep 30 14:10:34 crc kubenswrapper[4840]: I0930 14:10:34.869523 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-jwlj9" event={"ID":"5daf5662-8848-4fd9-a6af-e091fe303d89","Type":"ContainerStarted","Data":"ad2d0169fec767980eb85c49a78f751ff13758067e735f04fb3a383d2f7920cc"} Sep 30 14:10:34 crc kubenswrapper[4840]: I0930 14:10:34.869911 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-jwlj9" Sep 30 14:10:34 crc kubenswrapper[4840]: I0930 14:10:34.875611 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-x4mx4" podStartSLOduration=5.136836263 podStartE2EDuration="26.875594994s" podCreationTimestamp="2025-09-30 14:10:08 +0000 UTC" firstStartedPulling="2025-09-30 14:10:10.475406311 +0000 UTC m=+839.104492734" lastFinishedPulling="2025-09-30 14:10:32.214165042 +0000 UTC m=+860.843251465" observedRunningTime="2025-09-30 14:10:34.871050606 +0000 UTC m=+863.500137029" watchObservedRunningTime="2025-09-30 14:10:34.875594994 +0000 UTC m=+863.504681417" Sep 30 14:10:34 crc kubenswrapper[4840]: I0930 14:10:34.878852 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-xzmg8" event={"ID":"385365f0-9e2c-4c89-b95b-164a473ff0c3","Type":"ContainerStarted","Data":"7cad9b3a11b3197027d1e8e451cf6736e8369a1e96207f91e914ea4c8dcb38c0"} Sep 30 14:10:34 crc kubenswrapper[4840]: I0930 14:10:34.879157 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-88c7-xzmg8" Sep 30 14:10:34 crc kubenswrapper[4840]: I0930 14:10:34.895333 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-smg6p" event={"ID":"3ffb0dd0-3dd8-4f37-8ead-4add746a1fc6","Type":"ContainerStarted","Data":"92576c820a33de1bdef27a74de4b53740ec46977db1057d9a7334094328cc975"} Sep 30 14:10:34 crc kubenswrapper[4840]: I0930 14:10:34.895437 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-smg6p" Sep 30 14:10:34 crc kubenswrapper[4840]: I0930 14:10:34.910588 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-d2hgd" podStartSLOduration=4.982293804 podStartE2EDuration="26.910562793s" podCreationTimestamp="2025-09-30 14:10:08 +0000 UTC" firstStartedPulling="2025-09-30 14:10:10.294901187 +0000 UTC m=+838.923987610" lastFinishedPulling="2025-09-30 14:10:32.223170176 +0000 UTC m=+860.852256599" observedRunningTime="2025-09-30 14:10:34.895614365 +0000 UTC m=+863.524700788" watchObservedRunningTime="2025-09-30 14:10:34.910562793 +0000 UTC m=+863.539649216" Sep 30 14:10:34 crc kubenswrapper[4840]: I0930 14:10:34.918002 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-p6msl" event={"ID":"62475ec3-b883-4d16-bb31-6ee393062da2","Type":"ContainerStarted","Data":"70640a30d4102ef765c028a5206d88658a76dcba354810acbe252c898e49b521"} Sep 30 14:10:34 crc kubenswrapper[4840]: I0930 14:10:34.918301 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-p6msl" Sep 30 14:10:34 crc kubenswrapper[4840]: I0930 14:10:34.929433 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-fg7c7" event={"ID":"494a382c-6fec-4711-8a90-7f691c847e86","Type":"ContainerStarted","Data":"77ea99eebd9577d69f06d40ebb64c26251ec1925a7524dc1b2201a6159cd1252"} Sep 30 14:10:34 crc kubenswrapper[4840]: I0930 14:10:34.930697 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-fg7c7" Sep 30 14:10:34 crc kubenswrapper[4840]: I0930 14:10:34.940722 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-9t25m" podStartSLOduration=12.896476104 podStartE2EDuration="26.940697247s" podCreationTimestamp="2025-09-30 14:10:08 +0000 UTC" firstStartedPulling="2025-09-30 14:10:10.458462468 +0000 UTC m=+839.087548891" lastFinishedPulling="2025-09-30 14:10:24.502683591 +0000 UTC m=+853.131770034" observedRunningTime="2025-09-30 14:10:34.92887984 +0000 UTC m=+863.557966273" watchObservedRunningTime="2025-09-30 14:10:34.940697247 +0000 UTC m=+863.569783680" Sep 30 14:10:34 crc kubenswrapper[4840]: I0930 14:10:34.949885 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-4mpb8" event={"ID":"2b8703c9-f691-4c3a-9941-8bcd36aaf948","Type":"ContainerStarted","Data":"9bf0d7f2abe0f838f99dc5a987be216ccd8d34fd3e328323c8e7735d7ab03d52"} Sep 30 14:10:34 crc kubenswrapper[4840]: I0930 14:10:34.951313 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-4mpb8" Sep 30 14:10:34 crc kubenswrapper[4840]: I0930 14:10:34.952688 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-d9jgv" event={"ID":"63d94c36-0a55-472c-bbe8-27195105584e","Type":"ContainerStarted","Data":"ba46f65c528d658d28a97288de5f5a9bb38d7e1967ba04d690adf69d32831a51"} Sep 30 14:10:34 crc kubenswrapper[4840]: I0930 14:10:34.952921 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-d9jgv" Sep 30 14:10:34 crc kubenswrapper[4840]: I0930 14:10:34.955533 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-t2mx4" event={"ID":"a76ab4ec-b180-473f-a597-641725526b58","Type":"ContainerStarted","Data":"a99bf1a7de42dcf7ff0cd8539be80dd0d155bb112c8c867b884d60af6de89dac"} Sep 30 14:10:34 crc kubenswrapper[4840]: I0930 14:10:34.955982 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-t2mx4" Sep 30 14:10:34 crc kubenswrapper[4840]: I0930 14:10:34.962313 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-sq5hd" event={"ID":"64c5ad91-f113-46e3-ac87-36521d32f4f0","Type":"ContainerStarted","Data":"cbd37a1a1f23937d08e2df1ddb2e9f30e43877b6da32bd79f8f61189561360f3"} Sep 30 14:10:34 crc kubenswrapper[4840]: I0930 14:10:34.962516 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-sq5hd" Sep 30 14:10:34 crc kubenswrapper[4840]: I0930 14:10:34.967117 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-jwlj9" podStartSLOduration=5.039708771 podStartE2EDuration="26.967097093s" podCreationTimestamp="2025-09-30 14:10:08 +0000 UTC" firstStartedPulling="2025-09-30 14:10:10.286052661 +0000 UTC m=+838.915139084" lastFinishedPulling="2025-09-30 14:10:32.213440983 +0000 UTC m=+860.842527406" observedRunningTime="2025-09-30 14:10:34.952128714 +0000 UTC m=+863.581215157" watchObservedRunningTime="2025-09-30 14:10:34.967097093 +0000 UTC m=+863.596183536" Sep 30 14:10:34 crc kubenswrapper[4840]: I0930 14:10:34.967948 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-xtqwd" podStartSLOduration=5.616971222 podStartE2EDuration="26.967940725s" podCreationTimestamp="2025-09-30 14:10:08 +0000 UTC" firstStartedPulling="2025-09-30 14:10:10.872291325 +0000 UTC m=+839.501377738" lastFinishedPulling="2025-09-30 14:10:32.223260818 +0000 UTC m=+860.852347241" observedRunningTime="2025-09-30 14:10:34.965325067 +0000 UTC m=+863.594411490" watchObservedRunningTime="2025-09-30 14:10:34.967940725 +0000 UTC m=+863.597027158" Sep 30 14:10:35 crc kubenswrapper[4840]: I0930 14:10:35.007016 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-smg6p" podStartSLOduration=4.576819765 podStartE2EDuration="27.006999701s" podCreationTimestamp="2025-09-30 14:10:08 +0000 UTC" firstStartedPulling="2025-09-30 14:10:09.793153594 +0000 UTC m=+838.422240017" lastFinishedPulling="2025-09-30 14:10:32.22333353 +0000 UTC m=+860.852419953" observedRunningTime="2025-09-30 14:10:35.004672181 +0000 UTC m=+863.633758604" watchObservedRunningTime="2025-09-30 14:10:35.006999701 +0000 UTC m=+863.636086124" Sep 30 14:10:35 crc kubenswrapper[4840]: I0930 14:10:35.035097 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-p6msl" podStartSLOduration=5.688580839 podStartE2EDuration="27.035080091s" podCreationTimestamp="2025-09-30 14:10:08 +0000 UTC" firstStartedPulling="2025-09-30 14:10:10.883172463 +0000 UTC m=+839.512258886" lastFinishedPulling="2025-09-30 14:10:32.229671715 +0000 UTC m=+860.858758138" observedRunningTime="2025-09-30 14:10:35.032139305 +0000 UTC m=+863.661225728" watchObservedRunningTime="2025-09-30 14:10:35.035080091 +0000 UTC m=+863.664166504" Sep 30 14:10:35 crc kubenswrapper[4840]: I0930 14:10:35.054826 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-t2mx4" podStartSLOduration=5.289690295 podStartE2EDuration="27.054808074s" podCreationTimestamp="2025-09-30 14:10:08 +0000 UTC" firstStartedPulling="2025-09-30 14:10:10.458125079 +0000 UTC m=+839.087211502" lastFinishedPulling="2025-09-30 14:10:32.223242858 +0000 UTC m=+860.852329281" observedRunningTime="2025-09-30 14:10:35.048908591 +0000 UTC m=+863.677995014" watchObservedRunningTime="2025-09-30 14:10:35.054808074 +0000 UTC m=+863.683894487" Sep 30 14:10:35 crc kubenswrapper[4840]: I0930 14:10:35.068039 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-d9jgv" podStartSLOduration=12.85430781 podStartE2EDuration="27.068016858s" podCreationTimestamp="2025-09-30 14:10:08 +0000 UTC" firstStartedPulling="2025-09-30 14:10:10.294538028 +0000 UTC m=+838.923624451" lastFinishedPulling="2025-09-30 14:10:24.508247076 +0000 UTC m=+853.137333499" observedRunningTime="2025-09-30 14:10:35.063816738 +0000 UTC m=+863.692903161" watchObservedRunningTime="2025-09-30 14:10:35.068016858 +0000 UTC m=+863.697103281" Sep 30 14:10:35 crc kubenswrapper[4840]: I0930 14:10:35.097716 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-sq5hd" podStartSLOduration=12.841821877 podStartE2EDuration="27.097676519s" podCreationTimestamp="2025-09-30 14:10:08 +0000 UTC" firstStartedPulling="2025-09-30 14:10:10.247124637 +0000 UTC m=+838.876211060" lastFinishedPulling="2025-09-30 14:10:24.502979279 +0000 UTC m=+853.132065702" observedRunningTime="2025-09-30 14:10:35.082934085 +0000 UTC m=+863.712020518" watchObservedRunningTime="2025-09-30 14:10:35.097676519 +0000 UTC m=+863.726762942" Sep 30 14:10:35 crc kubenswrapper[4840]: I0930 14:10:35.099730 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-4mpb8" podStartSLOduration=4.150663302 podStartE2EDuration="26.099722962s" podCreationTimestamp="2025-09-30 14:10:09 +0000 UTC" firstStartedPulling="2025-09-30 14:10:10.27425665 +0000 UTC m=+838.903343073" lastFinishedPulling="2025-09-30 14:10:32.22331631 +0000 UTC m=+860.852402733" observedRunningTime="2025-09-30 14:10:35.097904695 +0000 UTC m=+863.726991118" watchObservedRunningTime="2025-09-30 14:10:35.099722962 +0000 UTC m=+863.728809385" Sep 30 14:10:35 crc kubenswrapper[4840]: I0930 14:10:35.136385 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-88c7-xzmg8" podStartSLOduration=5.204086924 podStartE2EDuration="27.136351994s" podCreationTimestamp="2025-09-30 14:10:08 +0000 UTC" firstStartedPulling="2025-09-30 14:10:10.288794231 +0000 UTC m=+838.917880654" lastFinishedPulling="2025-09-30 14:10:32.221059301 +0000 UTC m=+860.850145724" observedRunningTime="2025-09-30 14:10:35.1319549 +0000 UTC m=+863.761041333" watchObservedRunningTime="2025-09-30 14:10:35.136351994 +0000 UTC m=+863.765438417" Sep 30 14:10:35 crc kubenswrapper[4840]: I0930 14:10:35.136616 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-fg7c7" podStartSLOduration=12.895626092 podStartE2EDuration="27.136611561s" podCreationTimestamp="2025-09-30 14:10:08 +0000 UTC" firstStartedPulling="2025-09-30 14:10:10.256906827 +0000 UTC m=+838.885993240" lastFinishedPulling="2025-09-30 14:10:24.497892286 +0000 UTC m=+853.126978709" observedRunningTime="2025-09-30 14:10:35.117914655 +0000 UTC m=+863.747001078" watchObservedRunningTime="2025-09-30 14:10:35.136611561 +0000 UTC m=+863.765697984" Sep 30 14:10:38 crc kubenswrapper[4840]: I0930 14:10:38.922616 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-rnsbn" Sep 30 14:10:38 crc kubenswrapper[4840]: I0930 14:10:38.939910 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-smg6p" Sep 30 14:10:38 crc kubenswrapper[4840]: I0930 14:10:38.997640 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-f66b554c6-9npm5" event={"ID":"3ba6da78-9606-4e1c-bc55-6cbe2e052464","Type":"ContainerStarted","Data":"372c07c5eef4a44a88ca42d32517b6e143f8e6a5809ae04e9bce3f38f3dffaee"} Sep 30 14:10:38 crc kubenswrapper[4840]: I0930 14:10:38.997857 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-f66b554c6-9npm5" Sep 30 14:10:38 crc kubenswrapper[4840]: I0930 14:10:38.998718 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-6nn29" Sep 30 14:10:39 crc kubenswrapper[4840]: I0930 14:10:39.000202 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rzkrh" event={"ID":"0ae5abcf-aeb1-4607-8285-cd6ea1d1f912","Type":"ContainerStarted","Data":"634049030a2953f7e6844f650854d77d76d200739dc5797673406a90a50591c4"} Sep 30 14:10:39 crc kubenswrapper[4840]: I0930 14:10:39.002303 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-tqjvl" event={"ID":"3ed332eb-b604-4667-90f4-541d5a9115c2","Type":"ContainerStarted","Data":"dcc4628b82d724e4554d7831ea8e871d4dbe11d18843337c64deeb3ca9854f30"} Sep 30 14:10:39 crc kubenswrapper[4840]: I0930 14:10:39.002465 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-tqjvl" Sep 30 14:10:39 crc kubenswrapper[4840]: I0930 14:10:39.004067 4840 generic.go:334] "Generic (PLEG): container finished" podID="d3d4cfc5-2801-47a0-9c44-1943efdaa0ae" containerID="169b07113ba74ec95b72ec15530d61df08c7841dbb4775bd524d0e62831d7bb4" exitCode=0 Sep 30 14:10:39 crc kubenswrapper[4840]: I0930 14:10:39.004113 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-47pd9" event={"ID":"d3d4cfc5-2801-47a0-9c44-1943efdaa0ae","Type":"ContainerDied","Data":"169b07113ba74ec95b72ec15530d61df08c7841dbb4775bd524d0e62831d7bb4"} Sep 30 14:10:39 crc kubenswrapper[4840]: I0930 14:10:39.010855 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-smjg7" event={"ID":"aa2cf241-ec41-4a94-b452-653b3439a073","Type":"ContainerStarted","Data":"17ed0fff75f4396f4ae839e9436a1c1522047df1e05ecc3ec8cab80c67f82321"} Sep 30 14:10:39 crc kubenswrapper[4840]: I0930 14:10:39.015594 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-sq5hd" Sep 30 14:10:39 crc kubenswrapper[4840]: I0930 14:10:39.019694 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-qsq84" event={"ID":"3d5da746-e5f3-4bed-9c77-37a78da8f1a6","Type":"ContainerStarted","Data":"ea95f0e25af2205c06f5fb4d752285682f27f01d06a5b25c9e47b7a33895d25b"} Sep 30 14:10:39 crc kubenswrapper[4840]: I0930 14:10:39.020080 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-qsq84" Sep 30 14:10:39 crc kubenswrapper[4840]: I0930 14:10:39.022230 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-78974455b9-97rs4" event={"ID":"47248483-6556-491b-b65f-01cad9f1d1fc","Type":"ContainerStarted","Data":"d2bd0fc89560598c84eef7969231d8e78be4f42beda9e2e29e2266a54363fca8"} Sep 30 14:10:39 crc kubenswrapper[4840]: I0930 14:10:39.022465 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-f66b554c6-9npm5" podStartSLOduration=2.689007518 podStartE2EDuration="30.022454938s" podCreationTimestamp="2025-09-30 14:10:09 +0000 UTC" firstStartedPulling="2025-09-30 14:10:10.921630666 +0000 UTC m=+839.550717089" lastFinishedPulling="2025-09-30 14:10:38.255078086 +0000 UTC m=+866.884164509" observedRunningTime="2025-09-30 14:10:39.01985003 +0000 UTC m=+867.648936443" watchObservedRunningTime="2025-09-30 14:10:39.022454938 +0000 UTC m=+867.651541361" Sep 30 14:10:39 crc kubenswrapper[4840]: I0930 14:10:39.022852 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-78974455b9-97rs4" Sep 30 14:10:39 crc kubenswrapper[4840]: I0930 14:10:39.024677 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-rn2qq" event={"ID":"e1cd519e-b5c3-48f2-b426-d0f4e912683b","Type":"ContainerStarted","Data":"ab272205abc1840461b3d2be38b511db387fbf3e9322695734a6c4c93e42248f"} Sep 30 14:10:39 crc kubenswrapper[4840]: I0930 14:10:39.025136 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-rn2qq" Sep 30 14:10:39 crc kubenswrapper[4840]: I0930 14:10:39.044066 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-jwlj9" Sep 30 14:10:39 crc kubenswrapper[4840]: I0930 14:10:39.052218 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-d9jgv" Sep 30 14:10:39 crc kubenswrapper[4840]: I0930 14:10:39.060965 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rzkrh" podStartSLOduration=3.199372247 podStartE2EDuration="29.060950179s" podCreationTimestamp="2025-09-30 14:10:10 +0000 UTC" firstStartedPulling="2025-09-30 14:10:12.512730184 +0000 UTC m=+841.141816607" lastFinishedPulling="2025-09-30 14:10:38.374308116 +0000 UTC m=+867.003394539" observedRunningTime="2025-09-30 14:10:39.059425839 +0000 UTC m=+867.688512252" watchObservedRunningTime="2025-09-30 14:10:39.060950179 +0000 UTC m=+867.690036602" Sep 30 14:10:39 crc kubenswrapper[4840]: I0930 14:10:39.081224 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-tqjvl" podStartSLOduration=3.196327412 podStartE2EDuration="31.081206646s" podCreationTimestamp="2025-09-30 14:10:08 +0000 UTC" firstStartedPulling="2025-09-30 14:10:10.461473045 +0000 UTC m=+839.090559468" lastFinishedPulling="2025-09-30 14:10:38.346352279 +0000 UTC m=+866.975438702" observedRunningTime="2025-09-30 14:10:39.077360806 +0000 UTC m=+867.706447229" watchObservedRunningTime="2025-09-30 14:10:39.081206646 +0000 UTC m=+867.710293069" Sep 30 14:10:39 crc kubenswrapper[4840]: I0930 14:10:39.089304 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-smjg7" podStartSLOduration=2.6231623490000002 podStartE2EDuration="30.089290346s" podCreationTimestamp="2025-09-30 14:10:09 +0000 UTC" firstStartedPulling="2025-09-30 14:10:10.891936737 +0000 UTC m=+839.521023160" lastFinishedPulling="2025-09-30 14:10:38.358064724 +0000 UTC m=+866.987151157" observedRunningTime="2025-09-30 14:10:39.087296734 +0000 UTC m=+867.716383157" watchObservedRunningTime="2025-09-30 14:10:39.089290346 +0000 UTC m=+867.718376759" Sep 30 14:10:39 crc kubenswrapper[4840]: I0930 14:10:39.104746 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-qsq84" podStartSLOduration=3.326270513 podStartE2EDuration="31.104725337s" podCreationTimestamp="2025-09-30 14:10:08 +0000 UTC" firstStartedPulling="2025-09-30 14:10:10.475846112 +0000 UTC m=+839.104932535" lastFinishedPulling="2025-09-30 14:10:38.254300936 +0000 UTC m=+866.883387359" observedRunningTime="2025-09-30 14:10:39.101083823 +0000 UTC m=+867.730170246" watchObservedRunningTime="2025-09-30 14:10:39.104725337 +0000 UTC m=+867.733811760" Sep 30 14:10:39 crc kubenswrapper[4840]: I0930 14:10:39.160423 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-rn2qq" podStartSLOduration=2.830093766 podStartE2EDuration="30.160405445s" podCreationTimestamp="2025-09-30 14:10:09 +0000 UTC" firstStartedPulling="2025-09-30 14:10:10.923952236 +0000 UTC m=+839.553038659" lastFinishedPulling="2025-09-30 14:10:38.254263905 +0000 UTC m=+866.883350338" observedRunningTime="2025-09-30 14:10:39.15868882 +0000 UTC m=+867.787775263" watchObservedRunningTime="2025-09-30 14:10:39.160405445 +0000 UTC m=+867.789491868" Sep 30 14:10:39 crc kubenswrapper[4840]: I0930 14:10:39.175351 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-78974455b9-97rs4" podStartSLOduration=2.821675607 podStartE2EDuration="30.175334193s" podCreationTimestamp="2025-09-30 14:10:09 +0000 UTC" firstStartedPulling="2025-09-30 14:10:10.900960928 +0000 UTC m=+839.530047341" lastFinishedPulling="2025-09-30 14:10:38.254619504 +0000 UTC m=+866.883705927" observedRunningTime="2025-09-30 14:10:39.17214416 +0000 UTC m=+867.801230593" watchObservedRunningTime="2025-09-30 14:10:39.175334193 +0000 UTC m=+867.804420616" Sep 30 14:10:39 crc kubenswrapper[4840]: I0930 14:10:39.215742 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-fg7c7" Sep 30 14:10:39 crc kubenswrapper[4840]: I0930 14:10:39.228867 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-rjpbm" Sep 30 14:10:39 crc kubenswrapper[4840]: I0930 14:10:39.358152 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-4mpb8" Sep 30 14:10:39 crc kubenswrapper[4840]: I0930 14:10:39.365151 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-d2hgd" Sep 30 14:10:39 crc kubenswrapper[4840]: I0930 14:10:39.385407 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-88c7-xzmg8" Sep 30 14:10:39 crc kubenswrapper[4840]: I0930 14:10:39.430807 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-x4mx4" Sep 30 14:10:39 crc kubenswrapper[4840]: I0930 14:10:39.477093 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-t2mx4" Sep 30 14:10:39 crc kubenswrapper[4840]: I0930 14:10:39.595607 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-9t25m" Sep 30 14:10:39 crc kubenswrapper[4840]: I0930 14:10:39.799766 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-xtqwd" Sep 30 14:10:39 crc kubenswrapper[4840]: I0930 14:10:39.871843 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-p6msl" Sep 30 14:10:40 crc kubenswrapper[4840]: I0930 14:10:40.033468 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-47pd9" event={"ID":"d3d4cfc5-2801-47a0-9c44-1943efdaa0ae","Type":"ContainerStarted","Data":"266f37bdf1cabbc00a1c8517301db9bb43689548439110e3ead77ac85a293331"} Sep 30 14:10:40 crc kubenswrapper[4840]: I0930 14:10:40.065126 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-47pd9" podStartSLOduration=6.319990868 podStartE2EDuration="12.065104809s" podCreationTimestamp="2025-09-30 14:10:28 +0000 UTC" firstStartedPulling="2025-09-30 14:10:33.804117534 +0000 UTC m=+862.433203957" lastFinishedPulling="2025-09-30 14:10:39.549231475 +0000 UTC m=+868.178317898" observedRunningTime="2025-09-30 14:10:40.06129001 +0000 UTC m=+868.690376443" watchObservedRunningTime="2025-09-30 14:10:40.065104809 +0000 UTC m=+868.694191232" Sep 30 14:10:40 crc kubenswrapper[4840]: I0930 14:10:40.932257 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rzkrh" Sep 30 14:10:40 crc kubenswrapper[4840]: I0930 14:10:40.932311 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rzkrh" Sep 30 14:10:41 crc kubenswrapper[4840]: I0930 14:10:41.974964 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rzkrh" podUID="0ae5abcf-aeb1-4607-8285-cd6ea1d1f912" containerName="registry-server" probeResult="failure" output=< Sep 30 14:10:41 crc kubenswrapper[4840]: timeout: failed to connect service ":50051" within 1s Sep 30 14:10:41 crc kubenswrapper[4840]: > Sep 30 14:10:48 crc kubenswrapper[4840]: I0930 14:10:48.565387 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-47pd9" Sep 30 14:10:48 crc kubenswrapper[4840]: I0930 14:10:48.565726 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-47pd9" Sep 30 14:10:48 crc kubenswrapper[4840]: I0930 14:10:48.618907 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-47pd9" Sep 30 14:10:49 crc kubenswrapper[4840]: I0930 14:10:49.170800 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-47pd9" Sep 30 14:10:49 crc kubenswrapper[4840]: I0930 14:10:49.225956 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-47pd9"] Sep 30 14:10:49 crc kubenswrapper[4840]: I0930 14:10:49.536708 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-qsq84" Sep 30 14:10:49 crc kubenswrapper[4840]: I0930 14:10:49.624201 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-tqjvl" Sep 30 14:10:49 crc kubenswrapper[4840]: I0930 14:10:49.668611 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-78974455b9-97rs4" Sep 30 14:10:49 crc kubenswrapper[4840]: I0930 14:10:49.700146 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-f66b554c6-9npm5" Sep 30 14:10:49 crc kubenswrapper[4840]: I0930 14:10:49.724256 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-rn2qq" Sep 30 14:10:50 crc kubenswrapper[4840]: I0930 14:10:50.978777 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rzkrh" Sep 30 14:10:51 crc kubenswrapper[4840]: I0930 14:10:51.016688 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rzkrh" Sep 30 14:10:51 crc kubenswrapper[4840]: I0930 14:10:51.084164 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rzkrh"] Sep 30 14:10:51 crc kubenswrapper[4840]: I0930 14:10:51.119478 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-47pd9" podUID="d3d4cfc5-2801-47a0-9c44-1943efdaa0ae" containerName="registry-server" containerID="cri-o://266f37bdf1cabbc00a1c8517301db9bb43689548439110e3ead77ac85a293331" gracePeriod=2 Sep 30 14:10:51 crc kubenswrapper[4840]: I0930 14:10:51.252700 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4hhpc"] Sep 30 14:10:51 crc kubenswrapper[4840]: I0930 14:10:51.252962 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4hhpc" podUID="f2818028-42ff-459b-ab25-0f05227b88fd" containerName="registry-server" containerID="cri-o://ce79b7653d94d39c8257d5408f797f1a19beae2326136d5522df025cc846b0c4" gracePeriod=2 Sep 30 14:10:51 crc kubenswrapper[4840]: I0930 14:10:51.718338 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-47pd9" Sep 30 14:10:51 crc kubenswrapper[4840]: I0930 14:10:51.815395 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4hhpc" Sep 30 14:10:51 crc kubenswrapper[4840]: I0930 14:10:51.906522 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3d4cfc5-2801-47a0-9c44-1943efdaa0ae-utilities\") pod \"d3d4cfc5-2801-47a0-9c44-1943efdaa0ae\" (UID: \"d3d4cfc5-2801-47a0-9c44-1943efdaa0ae\") " Sep 30 14:10:51 crc kubenswrapper[4840]: I0930 14:10:51.906624 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3d4cfc5-2801-47a0-9c44-1943efdaa0ae-catalog-content\") pod \"d3d4cfc5-2801-47a0-9c44-1943efdaa0ae\" (UID: \"d3d4cfc5-2801-47a0-9c44-1943efdaa0ae\") " Sep 30 14:10:51 crc kubenswrapper[4840]: I0930 14:10:51.906695 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nntbs\" (UniqueName: \"kubernetes.io/projected/d3d4cfc5-2801-47a0-9c44-1943efdaa0ae-kube-api-access-nntbs\") pod \"d3d4cfc5-2801-47a0-9c44-1943efdaa0ae\" (UID: \"d3d4cfc5-2801-47a0-9c44-1943efdaa0ae\") " Sep 30 14:10:51 crc kubenswrapper[4840]: I0930 14:10:51.907535 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3d4cfc5-2801-47a0-9c44-1943efdaa0ae-utilities" (OuterVolumeSpecName: "utilities") pod "d3d4cfc5-2801-47a0-9c44-1943efdaa0ae" (UID: "d3d4cfc5-2801-47a0-9c44-1943efdaa0ae"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:10:51 crc kubenswrapper[4840]: I0930 14:10:51.912292 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3d4cfc5-2801-47a0-9c44-1943efdaa0ae-kube-api-access-nntbs" (OuterVolumeSpecName: "kube-api-access-nntbs") pod "d3d4cfc5-2801-47a0-9c44-1943efdaa0ae" (UID: "d3d4cfc5-2801-47a0-9c44-1943efdaa0ae"). InnerVolumeSpecName "kube-api-access-nntbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:10:51 crc kubenswrapper[4840]: I0930 14:10:51.918250 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3d4cfc5-2801-47a0-9c44-1943efdaa0ae-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d3d4cfc5-2801-47a0-9c44-1943efdaa0ae" (UID: "d3d4cfc5-2801-47a0-9c44-1943efdaa0ae"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.008028 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2818028-42ff-459b-ab25-0f05227b88fd-catalog-content\") pod \"f2818028-42ff-459b-ab25-0f05227b88fd\" (UID: \"f2818028-42ff-459b-ab25-0f05227b88fd\") " Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.008092 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l856r\" (UniqueName: \"kubernetes.io/projected/f2818028-42ff-459b-ab25-0f05227b88fd-kube-api-access-l856r\") pod \"f2818028-42ff-459b-ab25-0f05227b88fd\" (UID: \"f2818028-42ff-459b-ab25-0f05227b88fd\") " Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.008140 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2818028-42ff-459b-ab25-0f05227b88fd-utilities\") pod \"f2818028-42ff-459b-ab25-0f05227b88fd\" (UID: \"f2818028-42ff-459b-ab25-0f05227b88fd\") " Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.008450 4840 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3d4cfc5-2801-47a0-9c44-1943efdaa0ae-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.008467 4840 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3d4cfc5-2801-47a0-9c44-1943efdaa0ae-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.008481 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nntbs\" (UniqueName: \"kubernetes.io/projected/d3d4cfc5-2801-47a0-9c44-1943efdaa0ae-kube-api-access-nntbs\") on node \"crc\" DevicePath \"\"" Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.009692 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2818028-42ff-459b-ab25-0f05227b88fd-utilities" (OuterVolumeSpecName: "utilities") pod "f2818028-42ff-459b-ab25-0f05227b88fd" (UID: "f2818028-42ff-459b-ab25-0f05227b88fd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.011332 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2818028-42ff-459b-ab25-0f05227b88fd-kube-api-access-l856r" (OuterVolumeSpecName: "kube-api-access-l856r") pod "f2818028-42ff-459b-ab25-0f05227b88fd" (UID: "f2818028-42ff-459b-ab25-0f05227b88fd"). InnerVolumeSpecName "kube-api-access-l856r". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.095427 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2818028-42ff-459b-ab25-0f05227b88fd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f2818028-42ff-459b-ab25-0f05227b88fd" (UID: "f2818028-42ff-459b-ab25-0f05227b88fd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.109849 4840 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2818028-42ff-459b-ab25-0f05227b88fd-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.109875 4840 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2818028-42ff-459b-ab25-0f05227b88fd-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.109887 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l856r\" (UniqueName: \"kubernetes.io/projected/f2818028-42ff-459b-ab25-0f05227b88fd-kube-api-access-l856r\") on node \"crc\" DevicePath \"\"" Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.127007 4840 generic.go:334] "Generic (PLEG): container finished" podID="f2818028-42ff-459b-ab25-0f05227b88fd" containerID="ce79b7653d94d39c8257d5408f797f1a19beae2326136d5522df025cc846b0c4" exitCode=0 Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.127080 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4hhpc" event={"ID":"f2818028-42ff-459b-ab25-0f05227b88fd","Type":"ContainerDied","Data":"ce79b7653d94d39c8257d5408f797f1a19beae2326136d5522df025cc846b0c4"} Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.127103 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4hhpc" event={"ID":"f2818028-42ff-459b-ab25-0f05227b88fd","Type":"ContainerDied","Data":"cdd79db3e5dc511ee741e8d1ac32b6586a588a96981ffe6b25e900eb660221a4"} Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.127118 4840 scope.go:117] "RemoveContainer" containerID="ce79b7653d94d39c8257d5408f797f1a19beae2326136d5522df025cc846b0c4" Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.127108 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4hhpc" Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.129350 4840 generic.go:334] "Generic (PLEG): container finished" podID="d3d4cfc5-2801-47a0-9c44-1943efdaa0ae" containerID="266f37bdf1cabbc00a1c8517301db9bb43689548439110e3ead77ac85a293331" exitCode=0 Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.129484 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-47pd9" event={"ID":"d3d4cfc5-2801-47a0-9c44-1943efdaa0ae","Type":"ContainerDied","Data":"266f37bdf1cabbc00a1c8517301db9bb43689548439110e3ead77ac85a293331"} Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.129616 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-47pd9" event={"ID":"d3d4cfc5-2801-47a0-9c44-1943efdaa0ae","Type":"ContainerDied","Data":"7b73a3aa7e7eca3fb024fd1da37e42577e2cd4c262c0a8811c28068bb597acce"} Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.129508 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-47pd9" Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.148794 4840 scope.go:117] "RemoveContainer" containerID="007d9bb50d7a7033c1a985a9d5baf7f7a4ff2fe9aebd497b095cd58ba32c6592" Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.157285 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4hhpc"] Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.163484 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4hhpc"] Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.167896 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-47pd9"] Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.171794 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-47pd9"] Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.177562 4840 scope.go:117] "RemoveContainer" containerID="de0a31658503745458c591d53934c7bb41a9306b3a67314b29f088097f0e6c35" Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.192232 4840 scope.go:117] "RemoveContainer" containerID="ce79b7653d94d39c8257d5408f797f1a19beae2326136d5522df025cc846b0c4" Sep 30 14:10:52 crc kubenswrapper[4840]: E0930 14:10:52.192593 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce79b7653d94d39c8257d5408f797f1a19beae2326136d5522df025cc846b0c4\": container with ID starting with ce79b7653d94d39c8257d5408f797f1a19beae2326136d5522df025cc846b0c4 not found: ID does not exist" containerID="ce79b7653d94d39c8257d5408f797f1a19beae2326136d5522df025cc846b0c4" Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.192626 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce79b7653d94d39c8257d5408f797f1a19beae2326136d5522df025cc846b0c4"} err="failed to get container status \"ce79b7653d94d39c8257d5408f797f1a19beae2326136d5522df025cc846b0c4\": rpc error: code = NotFound desc = could not find container \"ce79b7653d94d39c8257d5408f797f1a19beae2326136d5522df025cc846b0c4\": container with ID starting with ce79b7653d94d39c8257d5408f797f1a19beae2326136d5522df025cc846b0c4 not found: ID does not exist" Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.192648 4840 scope.go:117] "RemoveContainer" containerID="007d9bb50d7a7033c1a985a9d5baf7f7a4ff2fe9aebd497b095cd58ba32c6592" Sep 30 14:10:52 crc kubenswrapper[4840]: E0930 14:10:52.192915 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"007d9bb50d7a7033c1a985a9d5baf7f7a4ff2fe9aebd497b095cd58ba32c6592\": container with ID starting with 007d9bb50d7a7033c1a985a9d5baf7f7a4ff2fe9aebd497b095cd58ba32c6592 not found: ID does not exist" containerID="007d9bb50d7a7033c1a985a9d5baf7f7a4ff2fe9aebd497b095cd58ba32c6592" Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.192941 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"007d9bb50d7a7033c1a985a9d5baf7f7a4ff2fe9aebd497b095cd58ba32c6592"} err="failed to get container status \"007d9bb50d7a7033c1a985a9d5baf7f7a4ff2fe9aebd497b095cd58ba32c6592\": rpc error: code = NotFound desc = could not find container \"007d9bb50d7a7033c1a985a9d5baf7f7a4ff2fe9aebd497b095cd58ba32c6592\": container with ID starting with 007d9bb50d7a7033c1a985a9d5baf7f7a4ff2fe9aebd497b095cd58ba32c6592 not found: ID does not exist" Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.192956 4840 scope.go:117] "RemoveContainer" containerID="de0a31658503745458c591d53934c7bb41a9306b3a67314b29f088097f0e6c35" Sep 30 14:10:52 crc kubenswrapper[4840]: E0930 14:10:52.193294 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de0a31658503745458c591d53934c7bb41a9306b3a67314b29f088097f0e6c35\": container with ID starting with de0a31658503745458c591d53934c7bb41a9306b3a67314b29f088097f0e6c35 not found: ID does not exist" containerID="de0a31658503745458c591d53934c7bb41a9306b3a67314b29f088097f0e6c35" Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.193319 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de0a31658503745458c591d53934c7bb41a9306b3a67314b29f088097f0e6c35"} err="failed to get container status \"de0a31658503745458c591d53934c7bb41a9306b3a67314b29f088097f0e6c35\": rpc error: code = NotFound desc = could not find container \"de0a31658503745458c591d53934c7bb41a9306b3a67314b29f088097f0e6c35\": container with ID starting with de0a31658503745458c591d53934c7bb41a9306b3a67314b29f088097f0e6c35 not found: ID does not exist" Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.193333 4840 scope.go:117] "RemoveContainer" containerID="266f37bdf1cabbc00a1c8517301db9bb43689548439110e3ead77ac85a293331" Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.215177 4840 scope.go:117] "RemoveContainer" containerID="169b07113ba74ec95b72ec15530d61df08c7841dbb4775bd524d0e62831d7bb4" Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.228500 4840 scope.go:117] "RemoveContainer" containerID="b1ab7febc7328622d487a292b00d0c98e61f871e6133ff2d908ca619c3ad25d8" Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.242654 4840 scope.go:117] "RemoveContainer" containerID="266f37bdf1cabbc00a1c8517301db9bb43689548439110e3ead77ac85a293331" Sep 30 14:10:52 crc kubenswrapper[4840]: E0930 14:10:52.243150 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"266f37bdf1cabbc00a1c8517301db9bb43689548439110e3ead77ac85a293331\": container with ID starting with 266f37bdf1cabbc00a1c8517301db9bb43689548439110e3ead77ac85a293331 not found: ID does not exist" containerID="266f37bdf1cabbc00a1c8517301db9bb43689548439110e3ead77ac85a293331" Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.243177 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"266f37bdf1cabbc00a1c8517301db9bb43689548439110e3ead77ac85a293331"} err="failed to get container status \"266f37bdf1cabbc00a1c8517301db9bb43689548439110e3ead77ac85a293331\": rpc error: code = NotFound desc = could not find container \"266f37bdf1cabbc00a1c8517301db9bb43689548439110e3ead77ac85a293331\": container with ID starting with 266f37bdf1cabbc00a1c8517301db9bb43689548439110e3ead77ac85a293331 not found: ID does not exist" Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.243197 4840 scope.go:117] "RemoveContainer" containerID="169b07113ba74ec95b72ec15530d61df08c7841dbb4775bd524d0e62831d7bb4" Sep 30 14:10:52 crc kubenswrapper[4840]: E0930 14:10:52.243445 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"169b07113ba74ec95b72ec15530d61df08c7841dbb4775bd524d0e62831d7bb4\": container with ID starting with 169b07113ba74ec95b72ec15530d61df08c7841dbb4775bd524d0e62831d7bb4 not found: ID does not exist" containerID="169b07113ba74ec95b72ec15530d61df08c7841dbb4775bd524d0e62831d7bb4" Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.243466 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"169b07113ba74ec95b72ec15530d61df08c7841dbb4775bd524d0e62831d7bb4"} err="failed to get container status \"169b07113ba74ec95b72ec15530d61df08c7841dbb4775bd524d0e62831d7bb4\": rpc error: code = NotFound desc = could not find container \"169b07113ba74ec95b72ec15530d61df08c7841dbb4775bd524d0e62831d7bb4\": container with ID starting with 169b07113ba74ec95b72ec15530d61df08c7841dbb4775bd524d0e62831d7bb4 not found: ID does not exist" Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.243479 4840 scope.go:117] "RemoveContainer" containerID="b1ab7febc7328622d487a292b00d0c98e61f871e6133ff2d908ca619c3ad25d8" Sep 30 14:10:52 crc kubenswrapper[4840]: E0930 14:10:52.243851 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1ab7febc7328622d487a292b00d0c98e61f871e6133ff2d908ca619c3ad25d8\": container with ID starting with b1ab7febc7328622d487a292b00d0c98e61f871e6133ff2d908ca619c3ad25d8 not found: ID does not exist" containerID="b1ab7febc7328622d487a292b00d0c98e61f871e6133ff2d908ca619c3ad25d8" Sep 30 14:10:52 crc kubenswrapper[4840]: I0930 14:10:52.243911 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1ab7febc7328622d487a292b00d0c98e61f871e6133ff2d908ca619c3ad25d8"} err="failed to get container status \"b1ab7febc7328622d487a292b00d0c98e61f871e6133ff2d908ca619c3ad25d8\": rpc error: code = NotFound desc = could not find container \"b1ab7febc7328622d487a292b00d0c98e61f871e6133ff2d908ca619c3ad25d8\": container with ID starting with b1ab7febc7328622d487a292b00d0c98e61f871e6133ff2d908ca619c3ad25d8 not found: ID does not exist" Sep 30 14:10:54 crc kubenswrapper[4840]: I0930 14:10:54.131764 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3d4cfc5-2801-47a0-9c44-1943efdaa0ae" path="/var/lib/kubelet/pods/d3d4cfc5-2801-47a0-9c44-1943efdaa0ae/volumes" Sep 30 14:10:54 crc kubenswrapper[4840]: I0930 14:10:54.133020 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2818028-42ff-459b-ab25-0f05227b88fd" path="/var/lib/kubelet/pods/f2818028-42ff-459b-ab25-0f05227b88fd/volumes" Sep 30 14:11:06 crc kubenswrapper[4840]: I0930 14:11:06.302743 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-7f6b5"] Sep 30 14:11:06 crc kubenswrapper[4840]: E0930 14:11:06.303367 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2818028-42ff-459b-ab25-0f05227b88fd" containerName="extract-content" Sep 30 14:11:06 crc kubenswrapper[4840]: I0930 14:11:06.303382 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2818028-42ff-459b-ab25-0f05227b88fd" containerName="extract-content" Sep 30 14:11:06 crc kubenswrapper[4840]: E0930 14:11:06.303420 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3d4cfc5-2801-47a0-9c44-1943efdaa0ae" containerName="extract-content" Sep 30 14:11:06 crc kubenswrapper[4840]: I0930 14:11:06.303429 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3d4cfc5-2801-47a0-9c44-1943efdaa0ae" containerName="extract-content" Sep 30 14:11:06 crc kubenswrapper[4840]: E0930 14:11:06.303446 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3d4cfc5-2801-47a0-9c44-1943efdaa0ae" containerName="registry-server" Sep 30 14:11:06 crc kubenswrapper[4840]: I0930 14:11:06.303455 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3d4cfc5-2801-47a0-9c44-1943efdaa0ae" containerName="registry-server" Sep 30 14:11:06 crc kubenswrapper[4840]: E0930 14:11:06.303473 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2818028-42ff-459b-ab25-0f05227b88fd" containerName="registry-server" Sep 30 14:11:06 crc kubenswrapper[4840]: I0930 14:11:06.303481 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2818028-42ff-459b-ab25-0f05227b88fd" containerName="registry-server" Sep 30 14:11:06 crc kubenswrapper[4840]: E0930 14:11:06.303504 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2818028-42ff-459b-ab25-0f05227b88fd" containerName="extract-utilities" Sep 30 14:11:06 crc kubenswrapper[4840]: I0930 14:11:06.303513 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2818028-42ff-459b-ab25-0f05227b88fd" containerName="extract-utilities" Sep 30 14:11:06 crc kubenswrapper[4840]: E0930 14:11:06.303525 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3d4cfc5-2801-47a0-9c44-1943efdaa0ae" containerName="extract-utilities" Sep 30 14:11:06 crc kubenswrapper[4840]: I0930 14:11:06.303533 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3d4cfc5-2801-47a0-9c44-1943efdaa0ae" containerName="extract-utilities" Sep 30 14:11:06 crc kubenswrapper[4840]: I0930 14:11:06.303719 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2818028-42ff-459b-ab25-0f05227b88fd" containerName="registry-server" Sep 30 14:11:06 crc kubenswrapper[4840]: I0930 14:11:06.303739 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3d4cfc5-2801-47a0-9c44-1943efdaa0ae" containerName="registry-server" Sep 30 14:11:06 crc kubenswrapper[4840]: I0930 14:11:06.304782 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-7f6b5" Sep 30 14:11:06 crc kubenswrapper[4840]: I0930 14:11:06.310090 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Sep 30 14:11:06 crc kubenswrapper[4840]: I0930 14:11:06.310178 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-58w52" Sep 30 14:11:06 crc kubenswrapper[4840]: I0930 14:11:06.310457 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Sep 30 14:11:06 crc kubenswrapper[4840]: I0930 14:11:06.310749 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Sep 30 14:11:06 crc kubenswrapper[4840]: I0930 14:11:06.317754 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-7f6b5"] Sep 30 14:11:06 crc kubenswrapper[4840]: I0930 14:11:06.353797 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-t9j4n"] Sep 30 14:11:06 crc kubenswrapper[4840]: I0930 14:11:06.355378 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-t9j4n" Sep 30 14:11:06 crc kubenswrapper[4840]: I0930 14:11:06.357395 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Sep 30 14:11:06 crc kubenswrapper[4840]: I0930 14:11:06.360282 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-t9j4n"] Sep 30 14:11:06 crc kubenswrapper[4840]: I0930 14:11:06.404634 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2-config\") pod \"dnsmasq-dns-78dd6ddcc-t9j4n\" (UID: \"39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-t9j4n" Sep 30 14:11:06 crc kubenswrapper[4840]: I0930 14:11:06.404685 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2m98\" (UniqueName: \"kubernetes.io/projected/39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2-kube-api-access-n2m98\") pod \"dnsmasq-dns-78dd6ddcc-t9j4n\" (UID: \"39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-t9j4n" Sep 30 14:11:06 crc kubenswrapper[4840]: I0930 14:11:06.404718 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgllj\" (UniqueName: \"kubernetes.io/projected/8230a9ad-c799-4e92-8841-31e87f8e22fb-kube-api-access-zgllj\") pod \"dnsmasq-dns-675f4bcbfc-7f6b5\" (UID: \"8230a9ad-c799-4e92-8841-31e87f8e22fb\") " pod="openstack/dnsmasq-dns-675f4bcbfc-7f6b5" Sep 30 14:11:06 crc kubenswrapper[4840]: I0930 14:11:06.404739 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-t9j4n\" (UID: \"39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-t9j4n" Sep 30 14:11:06 crc kubenswrapper[4840]: I0930 14:11:06.405087 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8230a9ad-c799-4e92-8841-31e87f8e22fb-config\") pod \"dnsmasq-dns-675f4bcbfc-7f6b5\" (UID: \"8230a9ad-c799-4e92-8841-31e87f8e22fb\") " pod="openstack/dnsmasq-dns-675f4bcbfc-7f6b5" Sep 30 14:11:06 crc kubenswrapper[4840]: I0930 14:11:06.505931 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2-config\") pod \"dnsmasq-dns-78dd6ddcc-t9j4n\" (UID: \"39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-t9j4n" Sep 30 14:11:06 crc kubenswrapper[4840]: I0930 14:11:06.505986 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2m98\" (UniqueName: \"kubernetes.io/projected/39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2-kube-api-access-n2m98\") pod \"dnsmasq-dns-78dd6ddcc-t9j4n\" (UID: \"39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-t9j4n" Sep 30 14:11:06 crc kubenswrapper[4840]: I0930 14:11:06.506020 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgllj\" (UniqueName: \"kubernetes.io/projected/8230a9ad-c799-4e92-8841-31e87f8e22fb-kube-api-access-zgllj\") pod \"dnsmasq-dns-675f4bcbfc-7f6b5\" (UID: \"8230a9ad-c799-4e92-8841-31e87f8e22fb\") " pod="openstack/dnsmasq-dns-675f4bcbfc-7f6b5" Sep 30 14:11:06 crc kubenswrapper[4840]: I0930 14:11:06.506043 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-t9j4n\" (UID: \"39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-t9j4n" Sep 30 14:11:06 crc kubenswrapper[4840]: I0930 14:11:06.506120 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8230a9ad-c799-4e92-8841-31e87f8e22fb-config\") pod \"dnsmasq-dns-675f4bcbfc-7f6b5\" (UID: \"8230a9ad-c799-4e92-8841-31e87f8e22fb\") " pod="openstack/dnsmasq-dns-675f4bcbfc-7f6b5" Sep 30 14:11:06 crc kubenswrapper[4840]: I0930 14:11:06.506889 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2-config\") pod \"dnsmasq-dns-78dd6ddcc-t9j4n\" (UID: \"39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-t9j4n" Sep 30 14:11:06 crc kubenswrapper[4840]: I0930 14:11:06.507046 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8230a9ad-c799-4e92-8841-31e87f8e22fb-config\") pod \"dnsmasq-dns-675f4bcbfc-7f6b5\" (UID: \"8230a9ad-c799-4e92-8841-31e87f8e22fb\") " pod="openstack/dnsmasq-dns-675f4bcbfc-7f6b5" Sep 30 14:11:06 crc kubenswrapper[4840]: I0930 14:11:06.507074 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-t9j4n\" (UID: \"39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-t9j4n" Sep 30 14:11:06 crc kubenswrapper[4840]: I0930 14:11:06.532301 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgllj\" (UniqueName: \"kubernetes.io/projected/8230a9ad-c799-4e92-8841-31e87f8e22fb-kube-api-access-zgllj\") pod \"dnsmasq-dns-675f4bcbfc-7f6b5\" (UID: \"8230a9ad-c799-4e92-8841-31e87f8e22fb\") " pod="openstack/dnsmasq-dns-675f4bcbfc-7f6b5" Sep 30 14:11:06 crc kubenswrapper[4840]: I0930 14:11:06.532320 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2m98\" (UniqueName: \"kubernetes.io/projected/39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2-kube-api-access-n2m98\") pod \"dnsmasq-dns-78dd6ddcc-t9j4n\" (UID: \"39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-t9j4n" Sep 30 14:11:06 crc kubenswrapper[4840]: I0930 14:11:06.636238 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-7f6b5" Sep 30 14:11:06 crc kubenswrapper[4840]: I0930 14:11:06.672005 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-t9j4n" Sep 30 14:11:06 crc kubenswrapper[4840]: I0930 14:11:06.907055 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-t9j4n"] Sep 30 14:11:07 crc kubenswrapper[4840]: I0930 14:11:07.033819 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-7f6b5"] Sep 30 14:11:07 crc kubenswrapper[4840]: W0930 14:11:07.039868 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8230a9ad_c799_4e92_8841_31e87f8e22fb.slice/crio-6dda08756fd4db73c84728907e67c65474ded83750972327a8bd47ea80db0ecc WatchSource:0}: Error finding container 6dda08756fd4db73c84728907e67c65474ded83750972327a8bd47ea80db0ecc: Status 404 returned error can't find the container with id 6dda08756fd4db73c84728907e67c65474ded83750972327a8bd47ea80db0ecc Sep 30 14:11:07 crc kubenswrapper[4840]: I0930 14:11:07.235775 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-t9j4n" event={"ID":"39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2","Type":"ContainerStarted","Data":"c623c300c431eb9ba4a2f0f8be3ff1e401ab2d8a40abb637a802d520e08117d1"} Sep 30 14:11:07 crc kubenswrapper[4840]: I0930 14:11:07.237038 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-7f6b5" event={"ID":"8230a9ad-c799-4e92-8841-31e87f8e22fb","Type":"ContainerStarted","Data":"6dda08756fd4db73c84728907e67c65474ded83750972327a8bd47ea80db0ecc"} Sep 30 14:11:09 crc kubenswrapper[4840]: I0930 14:11:09.157943 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-7f6b5"] Sep 30 14:11:09 crc kubenswrapper[4840]: I0930 14:11:09.188394 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-pvrll"] Sep 30 14:11:09 crc kubenswrapper[4840]: I0930 14:11:09.189782 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-pvrll" Sep 30 14:11:09 crc kubenswrapper[4840]: I0930 14:11:09.199889 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-pvrll"] Sep 30 14:11:09 crc kubenswrapper[4840]: I0930 14:11:09.259189 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qvr9\" (UniqueName: \"kubernetes.io/projected/9f2e45b7-fe3e-4e53-a735-05d843a14159-kube-api-access-9qvr9\") pod \"dnsmasq-dns-5ccc8479f9-pvrll\" (UID: \"9f2e45b7-fe3e-4e53-a735-05d843a14159\") " pod="openstack/dnsmasq-dns-5ccc8479f9-pvrll" Sep 30 14:11:09 crc kubenswrapper[4840]: I0930 14:11:09.259268 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f2e45b7-fe3e-4e53-a735-05d843a14159-config\") pod \"dnsmasq-dns-5ccc8479f9-pvrll\" (UID: \"9f2e45b7-fe3e-4e53-a735-05d843a14159\") " pod="openstack/dnsmasq-dns-5ccc8479f9-pvrll" Sep 30 14:11:09 crc kubenswrapper[4840]: I0930 14:11:09.259328 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f2e45b7-fe3e-4e53-a735-05d843a14159-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-pvrll\" (UID: \"9f2e45b7-fe3e-4e53-a735-05d843a14159\") " pod="openstack/dnsmasq-dns-5ccc8479f9-pvrll" Sep 30 14:11:09 crc kubenswrapper[4840]: I0930 14:11:09.359834 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qvr9\" (UniqueName: \"kubernetes.io/projected/9f2e45b7-fe3e-4e53-a735-05d843a14159-kube-api-access-9qvr9\") pod \"dnsmasq-dns-5ccc8479f9-pvrll\" (UID: \"9f2e45b7-fe3e-4e53-a735-05d843a14159\") " pod="openstack/dnsmasq-dns-5ccc8479f9-pvrll" Sep 30 14:11:09 crc kubenswrapper[4840]: I0930 14:11:09.359881 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f2e45b7-fe3e-4e53-a735-05d843a14159-config\") pod \"dnsmasq-dns-5ccc8479f9-pvrll\" (UID: \"9f2e45b7-fe3e-4e53-a735-05d843a14159\") " pod="openstack/dnsmasq-dns-5ccc8479f9-pvrll" Sep 30 14:11:09 crc kubenswrapper[4840]: I0930 14:11:09.359948 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f2e45b7-fe3e-4e53-a735-05d843a14159-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-pvrll\" (UID: \"9f2e45b7-fe3e-4e53-a735-05d843a14159\") " pod="openstack/dnsmasq-dns-5ccc8479f9-pvrll" Sep 30 14:11:09 crc kubenswrapper[4840]: I0930 14:11:09.361104 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f2e45b7-fe3e-4e53-a735-05d843a14159-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-pvrll\" (UID: \"9f2e45b7-fe3e-4e53-a735-05d843a14159\") " pod="openstack/dnsmasq-dns-5ccc8479f9-pvrll" Sep 30 14:11:09 crc kubenswrapper[4840]: I0930 14:11:09.361784 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f2e45b7-fe3e-4e53-a735-05d843a14159-config\") pod \"dnsmasq-dns-5ccc8479f9-pvrll\" (UID: \"9f2e45b7-fe3e-4e53-a735-05d843a14159\") " pod="openstack/dnsmasq-dns-5ccc8479f9-pvrll" Sep 30 14:11:09 crc kubenswrapper[4840]: I0930 14:11:09.391911 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qvr9\" (UniqueName: \"kubernetes.io/projected/9f2e45b7-fe3e-4e53-a735-05d843a14159-kube-api-access-9qvr9\") pod \"dnsmasq-dns-5ccc8479f9-pvrll\" (UID: \"9f2e45b7-fe3e-4e53-a735-05d843a14159\") " pod="openstack/dnsmasq-dns-5ccc8479f9-pvrll" Sep 30 14:11:09 crc kubenswrapper[4840]: I0930 14:11:09.416800 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-t9j4n"] Sep 30 14:11:09 crc kubenswrapper[4840]: I0930 14:11:09.438541 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-kb2gr"] Sep 30 14:11:09 crc kubenswrapper[4840]: I0930 14:11:09.439845 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-kb2gr" Sep 30 14:11:09 crc kubenswrapper[4840]: I0930 14:11:09.447087 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-kb2gr"] Sep 30 14:11:09 crc kubenswrapper[4840]: I0930 14:11:09.472997 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zg9qm\" (UniqueName: \"kubernetes.io/projected/67556acd-23a4-40ab-b0ba-32c267b52956-kube-api-access-zg9qm\") pod \"dnsmasq-dns-57d769cc4f-kb2gr\" (UID: \"67556acd-23a4-40ab-b0ba-32c267b52956\") " pod="openstack/dnsmasq-dns-57d769cc4f-kb2gr" Sep 30 14:11:09 crc kubenswrapper[4840]: I0930 14:11:09.473047 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67556acd-23a4-40ab-b0ba-32c267b52956-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-kb2gr\" (UID: \"67556acd-23a4-40ab-b0ba-32c267b52956\") " pod="openstack/dnsmasq-dns-57d769cc4f-kb2gr" Sep 30 14:11:09 crc kubenswrapper[4840]: I0930 14:11:09.473102 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67556acd-23a4-40ab-b0ba-32c267b52956-config\") pod \"dnsmasq-dns-57d769cc4f-kb2gr\" (UID: \"67556acd-23a4-40ab-b0ba-32c267b52956\") " pod="openstack/dnsmasq-dns-57d769cc4f-kb2gr" Sep 30 14:11:09 crc kubenswrapper[4840]: I0930 14:11:09.519762 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-pvrll" Sep 30 14:11:09 crc kubenswrapper[4840]: I0930 14:11:09.574985 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zg9qm\" (UniqueName: \"kubernetes.io/projected/67556acd-23a4-40ab-b0ba-32c267b52956-kube-api-access-zg9qm\") pod \"dnsmasq-dns-57d769cc4f-kb2gr\" (UID: \"67556acd-23a4-40ab-b0ba-32c267b52956\") " pod="openstack/dnsmasq-dns-57d769cc4f-kb2gr" Sep 30 14:11:09 crc kubenswrapper[4840]: I0930 14:11:09.575048 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67556acd-23a4-40ab-b0ba-32c267b52956-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-kb2gr\" (UID: \"67556acd-23a4-40ab-b0ba-32c267b52956\") " pod="openstack/dnsmasq-dns-57d769cc4f-kb2gr" Sep 30 14:11:09 crc kubenswrapper[4840]: I0930 14:11:09.575091 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67556acd-23a4-40ab-b0ba-32c267b52956-config\") pod \"dnsmasq-dns-57d769cc4f-kb2gr\" (UID: \"67556acd-23a4-40ab-b0ba-32c267b52956\") " pod="openstack/dnsmasq-dns-57d769cc4f-kb2gr" Sep 30 14:11:09 crc kubenswrapper[4840]: I0930 14:11:09.575963 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67556acd-23a4-40ab-b0ba-32c267b52956-config\") pod \"dnsmasq-dns-57d769cc4f-kb2gr\" (UID: \"67556acd-23a4-40ab-b0ba-32c267b52956\") " pod="openstack/dnsmasq-dns-57d769cc4f-kb2gr" Sep 30 14:11:09 crc kubenswrapper[4840]: I0930 14:11:09.577238 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67556acd-23a4-40ab-b0ba-32c267b52956-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-kb2gr\" (UID: \"67556acd-23a4-40ab-b0ba-32c267b52956\") " pod="openstack/dnsmasq-dns-57d769cc4f-kb2gr" Sep 30 14:11:09 crc kubenswrapper[4840]: I0930 14:11:09.602580 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zg9qm\" (UniqueName: \"kubernetes.io/projected/67556acd-23a4-40ab-b0ba-32c267b52956-kube-api-access-zg9qm\") pod \"dnsmasq-dns-57d769cc4f-kb2gr\" (UID: \"67556acd-23a4-40ab-b0ba-32c267b52956\") " pod="openstack/dnsmasq-dns-57d769cc4f-kb2gr" Sep 30 14:11:09 crc kubenswrapper[4840]: I0930 14:11:09.763874 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-kb2gr" Sep 30 14:11:09 crc kubenswrapper[4840]: I0930 14:11:09.982579 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-pvrll"] Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.319331 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.321103 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.327838 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.328075 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.328314 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-w66jp" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.328462 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.328605 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.328758 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.328898 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.336471 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.386486 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c15070b3-f247-4879-a7dc-618faf7e6e35-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.386573 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggjnw\" (UniqueName: \"kubernetes.io/projected/c15070b3-f247-4879-a7dc-618faf7e6e35-kube-api-access-ggjnw\") pod \"rabbitmq-cell1-server-0\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.386599 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.386630 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c15070b3-f247-4879-a7dc-618faf7e6e35-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.386653 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c15070b3-f247-4879-a7dc-618faf7e6e35-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.386684 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c15070b3-f247-4879-a7dc-618faf7e6e35-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.386712 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c15070b3-f247-4879-a7dc-618faf7e6e35-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.386978 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c15070b3-f247-4879-a7dc-618faf7e6e35-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.387103 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c15070b3-f247-4879-a7dc-618faf7e6e35-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.387144 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c15070b3-f247-4879-a7dc-618faf7e6e35-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.387307 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c15070b3-f247-4879-a7dc-618faf7e6e35-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.488863 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c15070b3-f247-4879-a7dc-618faf7e6e35-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.488935 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggjnw\" (UniqueName: \"kubernetes.io/projected/c15070b3-f247-4879-a7dc-618faf7e6e35-kube-api-access-ggjnw\") pod \"rabbitmq-cell1-server-0\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.488963 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.489015 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c15070b3-f247-4879-a7dc-618faf7e6e35-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.489039 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c15070b3-f247-4879-a7dc-618faf7e6e35-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.489074 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c15070b3-f247-4879-a7dc-618faf7e6e35-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.489106 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c15070b3-f247-4879-a7dc-618faf7e6e35-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.489133 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c15070b3-f247-4879-a7dc-618faf7e6e35-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.489158 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c15070b3-f247-4879-a7dc-618faf7e6e35-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.489179 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c15070b3-f247-4879-a7dc-618faf7e6e35-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.489211 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c15070b3-f247-4879-a7dc-618faf7e6e35-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.489428 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c15070b3-f247-4879-a7dc-618faf7e6e35-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.489581 4840 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.493116 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c15070b3-f247-4879-a7dc-618faf7e6e35-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.499526 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c15070b3-f247-4879-a7dc-618faf7e6e35-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.499642 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c15070b3-f247-4879-a7dc-618faf7e6e35-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.499829 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c15070b3-f247-4879-a7dc-618faf7e6e35-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.501593 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c15070b3-f247-4879-a7dc-618faf7e6e35-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.501784 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c15070b3-f247-4879-a7dc-618faf7e6e35-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.502059 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c15070b3-f247-4879-a7dc-618faf7e6e35-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.503928 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c15070b3-f247-4879-a7dc-618faf7e6e35-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.508862 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggjnw\" (UniqueName: \"kubernetes.io/projected/c15070b3-f247-4879-a7dc-618faf7e6e35-kube-api-access-ggjnw\") pod \"rabbitmq-cell1-server-0\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.513476 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.567107 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.568488 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.572493 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.572973 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.573193 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.573341 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.573390 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.573451 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-s4wjb" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.573617 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.578848 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.590263 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b8c84ed0-e20f-445e-9548-4ce979c23476-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " pod="openstack/rabbitmq-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.590310 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b8c84ed0-e20f-445e-9548-4ce979c23476-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " pod="openstack/rabbitmq-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.590344 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " pod="openstack/rabbitmq-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.590366 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b8c84ed0-e20f-445e-9548-4ce979c23476-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " pod="openstack/rabbitmq-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.590634 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b8c84ed0-e20f-445e-9548-4ce979c23476-config-data\") pod \"rabbitmq-server-0\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " pod="openstack/rabbitmq-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.590705 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b8c84ed0-e20f-445e-9548-4ce979c23476-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " pod="openstack/rabbitmq-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.590783 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b8c84ed0-e20f-445e-9548-4ce979c23476-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " pod="openstack/rabbitmq-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.590840 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b8c84ed0-e20f-445e-9548-4ce979c23476-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " pod="openstack/rabbitmq-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.590911 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pt89\" (UniqueName: \"kubernetes.io/projected/b8c84ed0-e20f-445e-9548-4ce979c23476-kube-api-access-5pt89\") pod \"rabbitmq-server-0\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " pod="openstack/rabbitmq-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.590991 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b8c84ed0-e20f-445e-9548-4ce979c23476-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " pod="openstack/rabbitmq-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.591036 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b8c84ed0-e20f-445e-9548-4ce979c23476-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " pod="openstack/rabbitmq-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.652254 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.691734 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b8c84ed0-e20f-445e-9548-4ce979c23476-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " pod="openstack/rabbitmq-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.691889 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b8c84ed0-e20f-445e-9548-4ce979c23476-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " pod="openstack/rabbitmq-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.691925 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b8c84ed0-e20f-445e-9548-4ce979c23476-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " pod="openstack/rabbitmq-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.691953 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pt89\" (UniqueName: \"kubernetes.io/projected/b8c84ed0-e20f-445e-9548-4ce979c23476-kube-api-access-5pt89\") pod \"rabbitmq-server-0\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " pod="openstack/rabbitmq-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.691994 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b8c84ed0-e20f-445e-9548-4ce979c23476-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " pod="openstack/rabbitmq-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.692023 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b8c84ed0-e20f-445e-9548-4ce979c23476-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " pod="openstack/rabbitmq-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.692057 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b8c84ed0-e20f-445e-9548-4ce979c23476-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " pod="openstack/rabbitmq-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.692080 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b8c84ed0-e20f-445e-9548-4ce979c23476-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " pod="openstack/rabbitmq-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.692112 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " pod="openstack/rabbitmq-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.692132 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b8c84ed0-e20f-445e-9548-4ce979c23476-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " pod="openstack/rabbitmq-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.692166 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b8c84ed0-e20f-445e-9548-4ce979c23476-config-data\") pod \"rabbitmq-server-0\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " pod="openstack/rabbitmq-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.692818 4840 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.693981 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b8c84ed0-e20f-445e-9548-4ce979c23476-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " pod="openstack/rabbitmq-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.694058 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b8c84ed0-e20f-445e-9548-4ce979c23476-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " pod="openstack/rabbitmq-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.693999 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b8c84ed0-e20f-445e-9548-4ce979c23476-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " pod="openstack/rabbitmq-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.694208 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b8c84ed0-e20f-445e-9548-4ce979c23476-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " pod="openstack/rabbitmq-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.694371 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b8c84ed0-e20f-445e-9548-4ce979c23476-config-data\") pod \"rabbitmq-server-0\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " pod="openstack/rabbitmq-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.696320 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b8c84ed0-e20f-445e-9548-4ce979c23476-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " pod="openstack/rabbitmq-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.697100 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b8c84ed0-e20f-445e-9548-4ce979c23476-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " pod="openstack/rabbitmq-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.697202 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b8c84ed0-e20f-445e-9548-4ce979c23476-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " pod="openstack/rabbitmq-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.699184 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b8c84ed0-e20f-445e-9548-4ce979c23476-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " pod="openstack/rabbitmq-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.710349 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pt89\" (UniqueName: \"kubernetes.io/projected/b8c84ed0-e20f-445e-9548-4ce979c23476-kube-api-access-5pt89\") pod \"rabbitmq-server-0\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " pod="openstack/rabbitmq-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.710429 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " pod="openstack/rabbitmq-server-0" Sep 30 14:11:10 crc kubenswrapper[4840]: I0930 14:11:10.898506 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 30 14:11:11 crc kubenswrapper[4840]: W0930 14:11:11.998021 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f2e45b7_fe3e_4e53_a735_05d843a14159.slice/crio-9fe96403f9dbc4f3c79eccdcece4ddf38c69fdec2c7499238253c979ed1b99bd WatchSource:0}: Error finding container 9fe96403f9dbc4f3c79eccdcece4ddf38c69fdec2c7499238253c979ed1b99bd: Status 404 returned error can't find the container with id 9fe96403f9dbc4f3c79eccdcece4ddf38c69fdec2c7499238253c979ed1b99bd Sep 30 14:11:12 crc kubenswrapper[4840]: I0930 14:11:12.001180 4840 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 14:11:12 crc kubenswrapper[4840]: I0930 14:11:12.290134 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-pvrll" event={"ID":"9f2e45b7-fe3e-4e53-a735-05d843a14159","Type":"ContainerStarted","Data":"9fe96403f9dbc4f3c79eccdcece4ddf38c69fdec2c7499238253c979ed1b99bd"} Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.334281 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.335786 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.341671 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.352665 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.353722 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-g9589" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.359490 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.362847 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.375983 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.377666 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.455542 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.458533 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.460655 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.461197 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.461283 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-5k7fx" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.462022 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.470541 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.531989 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f\") " pod="openstack/openstack-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.532052 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f-secrets\") pod \"openstack-galera-0\" (UID: \"9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f\") " pod="openstack/openstack-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.532085 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f\") " pod="openstack/openstack-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.532111 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f-config-data-generated\") pod \"openstack-galera-0\" (UID: \"9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f\") " pod="openstack/openstack-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.532135 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f-config-data-default\") pod \"openstack-galera-0\" (UID: \"9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f\") " pod="openstack/openstack-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.532149 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f-operator-scripts\") pod \"openstack-galera-0\" (UID: \"9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f\") " pod="openstack/openstack-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.532169 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f\") " pod="openstack/openstack-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.532186 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vll5\" (UniqueName: \"kubernetes.io/projected/9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f-kube-api-access-4vll5\") pod \"openstack-galera-0\" (UID: \"9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f\") " pod="openstack/openstack-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.532230 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f-kolla-config\") pod \"openstack-galera-0\" (UID: \"9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f\") " pod="openstack/openstack-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.633433 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/41b0d67b-4ef4-4ebb-b000-b37bbdac6b34-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"41b0d67b-4ef4-4ebb-b000-b37bbdac6b34\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.633485 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f\") " pod="openstack/openstack-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.633513 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f-config-data-generated\") pod \"openstack-galera-0\" (UID: \"9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f\") " pod="openstack/openstack-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.633538 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f-config-data-default\") pod \"openstack-galera-0\" (UID: \"9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f\") " pod="openstack/openstack-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.633580 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f-operator-scripts\") pod \"openstack-galera-0\" (UID: \"9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f\") " pod="openstack/openstack-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.633597 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41b0d67b-4ef4-4ebb-b000-b37bbdac6b34-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"41b0d67b-4ef4-4ebb-b000-b37bbdac6b34\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.633621 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41b0d67b-4ef4-4ebb-b000-b37bbdac6b34-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"41b0d67b-4ef4-4ebb-b000-b37bbdac6b34\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.633639 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f\") " pod="openstack/openstack-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.633660 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vll5\" (UniqueName: \"kubernetes.io/projected/9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f-kube-api-access-4vll5\") pod \"openstack-galera-0\" (UID: \"9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f\") " pod="openstack/openstack-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.633676 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/41b0d67b-4ef4-4ebb-b000-b37bbdac6b34-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"41b0d67b-4ef4-4ebb-b000-b37bbdac6b34\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.633702 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/41b0d67b-4ef4-4ebb-b000-b37bbdac6b34-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"41b0d67b-4ef4-4ebb-b000-b37bbdac6b34\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.633719 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gggjk\" (UniqueName: \"kubernetes.io/projected/41b0d67b-4ef4-4ebb-b000-b37bbdac6b34-kube-api-access-gggjk\") pod \"openstack-cell1-galera-0\" (UID: \"41b0d67b-4ef4-4ebb-b000-b37bbdac6b34\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.633760 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f-kolla-config\") pod \"openstack-galera-0\" (UID: \"9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f\") " pod="openstack/openstack-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.633782 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f\") " pod="openstack/openstack-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.633806 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/41b0d67b-4ef4-4ebb-b000-b37bbdac6b34-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"41b0d67b-4ef4-4ebb-b000-b37bbdac6b34\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.633825 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/41b0d67b-4ef4-4ebb-b000-b37bbdac6b34-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"41b0d67b-4ef4-4ebb-b000-b37bbdac6b34\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.633850 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"41b0d67b-4ef4-4ebb-b000-b37bbdac6b34\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.633868 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f-secrets\") pod \"openstack-galera-0\" (UID: \"9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f\") " pod="openstack/openstack-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.634041 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f-config-data-generated\") pod \"openstack-galera-0\" (UID: \"9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f\") " pod="openstack/openstack-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.634341 4840 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/openstack-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.635295 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f-config-data-default\") pod \"openstack-galera-0\" (UID: \"9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f\") " pod="openstack/openstack-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.635923 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f-kolla-config\") pod \"openstack-galera-0\" (UID: \"9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f\") " pod="openstack/openstack-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.636873 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f-operator-scripts\") pod \"openstack-galera-0\" (UID: \"9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f\") " pod="openstack/openstack-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.648357 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f\") " pod="openstack/openstack-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.648397 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f\") " pod="openstack/openstack-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.657184 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f-secrets\") pod \"openstack-galera-0\" (UID: \"9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f\") " pod="openstack/openstack-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.661424 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vll5\" (UniqueName: \"kubernetes.io/projected/9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f-kube-api-access-4vll5\") pod \"openstack-galera-0\" (UID: \"9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f\") " pod="openstack/openstack-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.670465 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f\") " pod="openstack/openstack-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.735184 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/41b0d67b-4ef4-4ebb-b000-b37bbdac6b34-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"41b0d67b-4ef4-4ebb-b000-b37bbdac6b34\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.735238 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gggjk\" (UniqueName: \"kubernetes.io/projected/41b0d67b-4ef4-4ebb-b000-b37bbdac6b34-kube-api-access-gggjk\") pod \"openstack-cell1-galera-0\" (UID: \"41b0d67b-4ef4-4ebb-b000-b37bbdac6b34\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.735312 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/41b0d67b-4ef4-4ebb-b000-b37bbdac6b34-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"41b0d67b-4ef4-4ebb-b000-b37bbdac6b34\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.735338 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/41b0d67b-4ef4-4ebb-b000-b37bbdac6b34-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"41b0d67b-4ef4-4ebb-b000-b37bbdac6b34\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.735370 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"41b0d67b-4ef4-4ebb-b000-b37bbdac6b34\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.735396 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/41b0d67b-4ef4-4ebb-b000-b37bbdac6b34-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"41b0d67b-4ef4-4ebb-b000-b37bbdac6b34\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.735450 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41b0d67b-4ef4-4ebb-b000-b37bbdac6b34-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"41b0d67b-4ef4-4ebb-b000-b37bbdac6b34\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.735477 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41b0d67b-4ef4-4ebb-b000-b37bbdac6b34-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"41b0d67b-4ef4-4ebb-b000-b37bbdac6b34\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.735511 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/41b0d67b-4ef4-4ebb-b000-b37bbdac6b34-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"41b0d67b-4ef4-4ebb-b000-b37bbdac6b34\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.736382 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/41b0d67b-4ef4-4ebb-b000-b37bbdac6b34-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"41b0d67b-4ef4-4ebb-b000-b37bbdac6b34\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.737145 4840 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"41b0d67b-4ef4-4ebb-b000-b37bbdac6b34\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/openstack-cell1-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.737229 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/41b0d67b-4ef4-4ebb-b000-b37bbdac6b34-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"41b0d67b-4ef4-4ebb-b000-b37bbdac6b34\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.741031 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/41b0d67b-4ef4-4ebb-b000-b37bbdac6b34-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"41b0d67b-4ef4-4ebb-b000-b37bbdac6b34\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.743256 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41b0d67b-4ef4-4ebb-b000-b37bbdac6b34-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"41b0d67b-4ef4-4ebb-b000-b37bbdac6b34\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.743274 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/41b0d67b-4ef4-4ebb-b000-b37bbdac6b34-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"41b0d67b-4ef4-4ebb-b000-b37bbdac6b34\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.747362 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/41b0d67b-4ef4-4ebb-b000-b37bbdac6b34-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"41b0d67b-4ef4-4ebb-b000-b37bbdac6b34\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.753420 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41b0d67b-4ef4-4ebb-b000-b37bbdac6b34-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"41b0d67b-4ef4-4ebb-b000-b37bbdac6b34\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.760441 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gggjk\" (UniqueName: \"kubernetes.io/projected/41b0d67b-4ef4-4ebb-b000-b37bbdac6b34-kube-api-access-gggjk\") pod \"openstack-cell1-galera-0\" (UID: \"41b0d67b-4ef4-4ebb-b000-b37bbdac6b34\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.774714 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"41b0d67b-4ef4-4ebb-b000-b37bbdac6b34\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.781958 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.835070 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.838514 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.840649 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.843248 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-4lhtw" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.844319 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.849098 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.937585 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79k6s\" (UniqueName: \"kubernetes.io/projected/f0207213-00d5-4b63-b041-fff191463f4a-kube-api-access-79k6s\") pod \"memcached-0\" (UID: \"f0207213-00d5-4b63-b041-fff191463f4a\") " pod="openstack/memcached-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.937622 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f0207213-00d5-4b63-b041-fff191463f4a-kolla-config\") pod \"memcached-0\" (UID: \"f0207213-00d5-4b63-b041-fff191463f4a\") " pod="openstack/memcached-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.937655 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0207213-00d5-4b63-b041-fff191463f4a-combined-ca-bundle\") pod \"memcached-0\" (UID: \"f0207213-00d5-4b63-b041-fff191463f4a\") " pod="openstack/memcached-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.937746 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f0207213-00d5-4b63-b041-fff191463f4a-config-data\") pod \"memcached-0\" (UID: \"f0207213-00d5-4b63-b041-fff191463f4a\") " pod="openstack/memcached-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.937773 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0207213-00d5-4b63-b041-fff191463f4a-memcached-tls-certs\") pod \"memcached-0\" (UID: \"f0207213-00d5-4b63-b041-fff191463f4a\") " pod="openstack/memcached-0" Sep 30 14:11:13 crc kubenswrapper[4840]: I0930 14:11:13.963229 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Sep 30 14:11:14 crc kubenswrapper[4840]: I0930 14:11:14.039028 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f0207213-00d5-4b63-b041-fff191463f4a-config-data\") pod \"memcached-0\" (UID: \"f0207213-00d5-4b63-b041-fff191463f4a\") " pod="openstack/memcached-0" Sep 30 14:11:14 crc kubenswrapper[4840]: I0930 14:11:14.039823 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0207213-00d5-4b63-b041-fff191463f4a-memcached-tls-certs\") pod \"memcached-0\" (UID: \"f0207213-00d5-4b63-b041-fff191463f4a\") " pod="openstack/memcached-0" Sep 30 14:11:14 crc kubenswrapper[4840]: I0930 14:11:14.039887 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f0207213-00d5-4b63-b041-fff191463f4a-config-data\") pod \"memcached-0\" (UID: \"f0207213-00d5-4b63-b041-fff191463f4a\") " pod="openstack/memcached-0" Sep 30 14:11:14 crc kubenswrapper[4840]: I0930 14:11:14.039954 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79k6s\" (UniqueName: \"kubernetes.io/projected/f0207213-00d5-4b63-b041-fff191463f4a-kube-api-access-79k6s\") pod \"memcached-0\" (UID: \"f0207213-00d5-4b63-b041-fff191463f4a\") " pod="openstack/memcached-0" Sep 30 14:11:14 crc kubenswrapper[4840]: I0930 14:11:14.039985 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f0207213-00d5-4b63-b041-fff191463f4a-kolla-config\") pod \"memcached-0\" (UID: \"f0207213-00d5-4b63-b041-fff191463f4a\") " pod="openstack/memcached-0" Sep 30 14:11:14 crc kubenswrapper[4840]: I0930 14:11:14.040028 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0207213-00d5-4b63-b041-fff191463f4a-combined-ca-bundle\") pod \"memcached-0\" (UID: \"f0207213-00d5-4b63-b041-fff191463f4a\") " pod="openstack/memcached-0" Sep 30 14:11:14 crc kubenswrapper[4840]: I0930 14:11:14.040653 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f0207213-00d5-4b63-b041-fff191463f4a-kolla-config\") pod \"memcached-0\" (UID: \"f0207213-00d5-4b63-b041-fff191463f4a\") " pod="openstack/memcached-0" Sep 30 14:11:14 crc kubenswrapper[4840]: I0930 14:11:14.042918 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0207213-00d5-4b63-b041-fff191463f4a-memcached-tls-certs\") pod \"memcached-0\" (UID: \"f0207213-00d5-4b63-b041-fff191463f4a\") " pod="openstack/memcached-0" Sep 30 14:11:14 crc kubenswrapper[4840]: I0930 14:11:14.057900 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0207213-00d5-4b63-b041-fff191463f4a-combined-ca-bundle\") pod \"memcached-0\" (UID: \"f0207213-00d5-4b63-b041-fff191463f4a\") " pod="openstack/memcached-0" Sep 30 14:11:14 crc kubenswrapper[4840]: I0930 14:11:14.059314 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79k6s\" (UniqueName: \"kubernetes.io/projected/f0207213-00d5-4b63-b041-fff191463f4a-kube-api-access-79k6s\") pod \"memcached-0\" (UID: \"f0207213-00d5-4b63-b041-fff191463f4a\") " pod="openstack/memcached-0" Sep 30 14:11:14 crc kubenswrapper[4840]: I0930 14:11:14.166187 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Sep 30 14:11:15 crc kubenswrapper[4840]: I0930 14:11:15.890830 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Sep 30 14:11:15 crc kubenswrapper[4840]: I0930 14:11:15.892408 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 30 14:11:15 crc kubenswrapper[4840]: I0930 14:11:15.895632 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-8p2bw" Sep 30 14:11:15 crc kubenswrapper[4840]: I0930 14:11:15.904993 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 30 14:11:16 crc kubenswrapper[4840]: I0930 14:11:16.071732 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7mxd\" (UniqueName: \"kubernetes.io/projected/3e51f892-d5b7-48b6-9a5f-0c0af8a00f21-kube-api-access-w7mxd\") pod \"kube-state-metrics-0\" (UID: \"3e51f892-d5b7-48b6-9a5f-0c0af8a00f21\") " pod="openstack/kube-state-metrics-0" Sep 30 14:11:16 crc kubenswrapper[4840]: I0930 14:11:16.173466 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7mxd\" (UniqueName: \"kubernetes.io/projected/3e51f892-d5b7-48b6-9a5f-0c0af8a00f21-kube-api-access-w7mxd\") pod \"kube-state-metrics-0\" (UID: \"3e51f892-d5b7-48b6-9a5f-0c0af8a00f21\") " pod="openstack/kube-state-metrics-0" Sep 30 14:11:16 crc kubenswrapper[4840]: I0930 14:11:16.193490 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7mxd\" (UniqueName: \"kubernetes.io/projected/3e51f892-d5b7-48b6-9a5f-0c0af8a00f21-kube-api-access-w7mxd\") pod \"kube-state-metrics-0\" (UID: \"3e51f892-d5b7-48b6-9a5f-0c0af8a00f21\") " pod="openstack/kube-state-metrics-0" Sep 30 14:11:16 crc kubenswrapper[4840]: I0930 14:11:16.211177 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 30 14:11:19 crc kubenswrapper[4840]: I0930 14:11:19.766175 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Sep 30 14:11:19 crc kubenswrapper[4840]: I0930 14:11:19.767328 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Sep 30 14:11:19 crc kubenswrapper[4840]: I0930 14:11:19.771681 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Sep 30 14:11:19 crc kubenswrapper[4840]: I0930 14:11:19.772447 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Sep 30 14:11:19 crc kubenswrapper[4840]: I0930 14:11:19.775169 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Sep 30 14:11:19 crc kubenswrapper[4840]: I0930 14:11:19.775311 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Sep 30 14:11:19 crc kubenswrapper[4840]: I0930 14:11:19.775428 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-6pbst" Sep 30 14:11:19 crc kubenswrapper[4840]: I0930 14:11:19.782752 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Sep 30 14:11:19 crc kubenswrapper[4840]: I0930 14:11:19.893207 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-cngvz"] Sep 30 14:11:19 crc kubenswrapper[4840]: I0930 14:11:19.894794 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-cngvz" Sep 30 14:11:19 crc kubenswrapper[4840]: I0930 14:11:19.899668 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Sep 30 14:11:19 crc kubenswrapper[4840]: I0930 14:11:19.899925 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-pbpj9" Sep 30 14:11:19 crc kubenswrapper[4840]: I0930 14:11:19.901817 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Sep 30 14:11:19 crc kubenswrapper[4840]: I0930 14:11:19.901970 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-cngvz"] Sep 30 14:11:19 crc kubenswrapper[4840]: I0930 14:11:19.911858 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-fc5jg"] Sep 30 14:11:19 crc kubenswrapper[4840]: I0930 14:11:19.915250 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-fc5jg" Sep 30 14:11:19 crc kubenswrapper[4840]: I0930 14:11:19.921399 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-fc5jg"] Sep 30 14:11:19 crc kubenswrapper[4840]: I0930 14:11:19.936642 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2082b33d-3b6c-449a-9bdd-63d665a3d313-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"2082b33d-3b6c-449a-9bdd-63d665a3d313\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:11:19 crc kubenswrapper[4840]: I0930 14:11:19.936696 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2082b33d-3b6c-449a-9bdd-63d665a3d313-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"2082b33d-3b6c-449a-9bdd-63d665a3d313\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:11:19 crc kubenswrapper[4840]: I0930 14:11:19.936740 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"2082b33d-3b6c-449a-9bdd-63d665a3d313\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:11:19 crc kubenswrapper[4840]: I0930 14:11:19.936797 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2082b33d-3b6c-449a-9bdd-63d665a3d313-config\") pod \"ovsdbserver-nb-0\" (UID: \"2082b33d-3b6c-449a-9bdd-63d665a3d313\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:11:19 crc kubenswrapper[4840]: I0930 14:11:19.936834 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flzm4\" (UniqueName: \"kubernetes.io/projected/2082b33d-3b6c-449a-9bdd-63d665a3d313-kube-api-access-flzm4\") pod \"ovsdbserver-nb-0\" (UID: \"2082b33d-3b6c-449a-9bdd-63d665a3d313\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:11:19 crc kubenswrapper[4840]: I0930 14:11:19.936884 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2082b33d-3b6c-449a-9bdd-63d665a3d313-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"2082b33d-3b6c-449a-9bdd-63d665a3d313\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:11:19 crc kubenswrapper[4840]: I0930 14:11:19.936922 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2082b33d-3b6c-449a-9bdd-63d665a3d313-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"2082b33d-3b6c-449a-9bdd-63d665a3d313\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:11:19 crc kubenswrapper[4840]: I0930 14:11:19.936972 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2082b33d-3b6c-449a-9bdd-63d665a3d313-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"2082b33d-3b6c-449a-9bdd-63d665a3d313\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.038312 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2082b33d-3b6c-449a-9bdd-63d665a3d313-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"2082b33d-3b6c-449a-9bdd-63d665a3d313\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.038376 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2082b33d-3b6c-449a-9bdd-63d665a3d313-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"2082b33d-3b6c-449a-9bdd-63d665a3d313\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.038403 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/7d1f7d84-d053-4c95-83ac-ecbff773463d-var-lib\") pod \"ovn-controller-ovs-fc5jg\" (UID: \"7d1f7d84-d053-4c95-83ac-ecbff773463d\") " pod="openstack/ovn-controller-ovs-fc5jg" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.038424 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6vsh\" (UniqueName: \"kubernetes.io/projected/7d1f7d84-d053-4c95-83ac-ecbff773463d-kube-api-access-n6vsh\") pod \"ovn-controller-ovs-fc5jg\" (UID: \"7d1f7d84-d053-4c95-83ac-ecbff773463d\") " pod="openstack/ovn-controller-ovs-fc5jg" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.038444 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8a63b50-ecd5-4993-a890-8c94bc2d5e60-combined-ca-bundle\") pod \"ovn-controller-cngvz\" (UID: \"e8a63b50-ecd5-4993-a890-8c94bc2d5e60\") " pod="openstack/ovn-controller-cngvz" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.038468 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e8a63b50-ecd5-4993-a890-8c94bc2d5e60-var-run-ovn\") pod \"ovn-controller-cngvz\" (UID: \"e8a63b50-ecd5-4993-a890-8c94bc2d5e60\") " pod="openstack/ovn-controller-cngvz" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.038502 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"2082b33d-3b6c-449a-9bdd-63d665a3d313\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.038532 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbdwf\" (UniqueName: \"kubernetes.io/projected/e8a63b50-ecd5-4993-a890-8c94bc2d5e60-kube-api-access-hbdwf\") pod \"ovn-controller-cngvz\" (UID: \"e8a63b50-ecd5-4993-a890-8c94bc2d5e60\") " pod="openstack/ovn-controller-cngvz" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.038572 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2082b33d-3b6c-449a-9bdd-63d665a3d313-config\") pod \"ovsdbserver-nb-0\" (UID: \"2082b33d-3b6c-449a-9bdd-63d665a3d313\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.038596 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/7d1f7d84-d053-4c95-83ac-ecbff773463d-var-log\") pod \"ovn-controller-ovs-fc5jg\" (UID: \"7d1f7d84-d053-4c95-83ac-ecbff773463d\") " pod="openstack/ovn-controller-ovs-fc5jg" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.038624 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flzm4\" (UniqueName: \"kubernetes.io/projected/2082b33d-3b6c-449a-9bdd-63d665a3d313-kube-api-access-flzm4\") pod \"ovsdbserver-nb-0\" (UID: \"2082b33d-3b6c-449a-9bdd-63d665a3d313\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.038663 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2082b33d-3b6c-449a-9bdd-63d665a3d313-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"2082b33d-3b6c-449a-9bdd-63d665a3d313\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.038681 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e8a63b50-ecd5-4993-a890-8c94bc2d5e60-var-log-ovn\") pod \"ovn-controller-cngvz\" (UID: \"e8a63b50-ecd5-4993-a890-8c94bc2d5e60\") " pod="openstack/ovn-controller-cngvz" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.038706 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7d1f7d84-d053-4c95-83ac-ecbff773463d-var-run\") pod \"ovn-controller-ovs-fc5jg\" (UID: \"7d1f7d84-d053-4c95-83ac-ecbff773463d\") " pod="openstack/ovn-controller-ovs-fc5jg" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.038731 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2082b33d-3b6c-449a-9bdd-63d665a3d313-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"2082b33d-3b6c-449a-9bdd-63d665a3d313\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.038748 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7d1f7d84-d053-4c95-83ac-ecbff773463d-scripts\") pod \"ovn-controller-ovs-fc5jg\" (UID: \"7d1f7d84-d053-4c95-83ac-ecbff773463d\") " pod="openstack/ovn-controller-ovs-fc5jg" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.038773 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e8a63b50-ecd5-4993-a890-8c94bc2d5e60-var-run\") pod \"ovn-controller-cngvz\" (UID: \"e8a63b50-ecd5-4993-a890-8c94bc2d5e60\") " pod="openstack/ovn-controller-cngvz" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.038812 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2082b33d-3b6c-449a-9bdd-63d665a3d313-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"2082b33d-3b6c-449a-9bdd-63d665a3d313\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.038838 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/7d1f7d84-d053-4c95-83ac-ecbff773463d-etc-ovs\") pod \"ovn-controller-ovs-fc5jg\" (UID: \"7d1f7d84-d053-4c95-83ac-ecbff773463d\") " pod="openstack/ovn-controller-ovs-fc5jg" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.038871 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e8a63b50-ecd5-4993-a890-8c94bc2d5e60-scripts\") pod \"ovn-controller-cngvz\" (UID: \"e8a63b50-ecd5-4993-a890-8c94bc2d5e60\") " pod="openstack/ovn-controller-cngvz" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.038896 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8a63b50-ecd5-4993-a890-8c94bc2d5e60-ovn-controller-tls-certs\") pod \"ovn-controller-cngvz\" (UID: \"e8a63b50-ecd5-4993-a890-8c94bc2d5e60\") " pod="openstack/ovn-controller-cngvz" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.039542 4840 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"2082b33d-3b6c-449a-9bdd-63d665a3d313\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/ovsdbserver-nb-0" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.040171 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2082b33d-3b6c-449a-9bdd-63d665a3d313-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"2082b33d-3b6c-449a-9bdd-63d665a3d313\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.040244 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2082b33d-3b6c-449a-9bdd-63d665a3d313-config\") pod \"ovsdbserver-nb-0\" (UID: \"2082b33d-3b6c-449a-9bdd-63d665a3d313\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.040431 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2082b33d-3b6c-449a-9bdd-63d665a3d313-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"2082b33d-3b6c-449a-9bdd-63d665a3d313\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.044001 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2082b33d-3b6c-449a-9bdd-63d665a3d313-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"2082b33d-3b6c-449a-9bdd-63d665a3d313\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.044029 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2082b33d-3b6c-449a-9bdd-63d665a3d313-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"2082b33d-3b6c-449a-9bdd-63d665a3d313\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.046039 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2082b33d-3b6c-449a-9bdd-63d665a3d313-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"2082b33d-3b6c-449a-9bdd-63d665a3d313\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.057573 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flzm4\" (UniqueName: \"kubernetes.io/projected/2082b33d-3b6c-449a-9bdd-63d665a3d313-kube-api-access-flzm4\") pod \"ovsdbserver-nb-0\" (UID: \"2082b33d-3b6c-449a-9bdd-63d665a3d313\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.070731 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"2082b33d-3b6c-449a-9bdd-63d665a3d313\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.095900 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.140830 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e8a63b50-ecd5-4993-a890-8c94bc2d5e60-var-log-ovn\") pod \"ovn-controller-cngvz\" (UID: \"e8a63b50-ecd5-4993-a890-8c94bc2d5e60\") " pod="openstack/ovn-controller-cngvz" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.140892 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7d1f7d84-d053-4c95-83ac-ecbff773463d-var-run\") pod \"ovn-controller-ovs-fc5jg\" (UID: \"7d1f7d84-d053-4c95-83ac-ecbff773463d\") " pod="openstack/ovn-controller-ovs-fc5jg" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.140920 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7d1f7d84-d053-4c95-83ac-ecbff773463d-scripts\") pod \"ovn-controller-ovs-fc5jg\" (UID: \"7d1f7d84-d053-4c95-83ac-ecbff773463d\") " pod="openstack/ovn-controller-ovs-fc5jg" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.140945 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e8a63b50-ecd5-4993-a890-8c94bc2d5e60-var-run\") pod \"ovn-controller-cngvz\" (UID: \"e8a63b50-ecd5-4993-a890-8c94bc2d5e60\") " pod="openstack/ovn-controller-cngvz" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.140977 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/7d1f7d84-d053-4c95-83ac-ecbff773463d-etc-ovs\") pod \"ovn-controller-ovs-fc5jg\" (UID: \"7d1f7d84-d053-4c95-83ac-ecbff773463d\") " pod="openstack/ovn-controller-ovs-fc5jg" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.141004 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e8a63b50-ecd5-4993-a890-8c94bc2d5e60-scripts\") pod \"ovn-controller-cngvz\" (UID: \"e8a63b50-ecd5-4993-a890-8c94bc2d5e60\") " pod="openstack/ovn-controller-cngvz" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.141020 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8a63b50-ecd5-4993-a890-8c94bc2d5e60-ovn-controller-tls-certs\") pod \"ovn-controller-cngvz\" (UID: \"e8a63b50-ecd5-4993-a890-8c94bc2d5e60\") " pod="openstack/ovn-controller-cngvz" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.141042 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/7d1f7d84-d053-4c95-83ac-ecbff773463d-var-lib\") pod \"ovn-controller-ovs-fc5jg\" (UID: \"7d1f7d84-d053-4c95-83ac-ecbff773463d\") " pod="openstack/ovn-controller-ovs-fc5jg" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.141062 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6vsh\" (UniqueName: \"kubernetes.io/projected/7d1f7d84-d053-4c95-83ac-ecbff773463d-kube-api-access-n6vsh\") pod \"ovn-controller-ovs-fc5jg\" (UID: \"7d1f7d84-d053-4c95-83ac-ecbff773463d\") " pod="openstack/ovn-controller-ovs-fc5jg" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.141080 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8a63b50-ecd5-4993-a890-8c94bc2d5e60-combined-ca-bundle\") pod \"ovn-controller-cngvz\" (UID: \"e8a63b50-ecd5-4993-a890-8c94bc2d5e60\") " pod="openstack/ovn-controller-cngvz" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.141097 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e8a63b50-ecd5-4993-a890-8c94bc2d5e60-var-run-ovn\") pod \"ovn-controller-cngvz\" (UID: \"e8a63b50-ecd5-4993-a890-8c94bc2d5e60\") " pod="openstack/ovn-controller-cngvz" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.141127 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbdwf\" (UniqueName: \"kubernetes.io/projected/e8a63b50-ecd5-4993-a890-8c94bc2d5e60-kube-api-access-hbdwf\") pod \"ovn-controller-cngvz\" (UID: \"e8a63b50-ecd5-4993-a890-8c94bc2d5e60\") " pod="openstack/ovn-controller-cngvz" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.141145 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/7d1f7d84-d053-4c95-83ac-ecbff773463d-var-log\") pod \"ovn-controller-ovs-fc5jg\" (UID: \"7d1f7d84-d053-4c95-83ac-ecbff773463d\") " pod="openstack/ovn-controller-ovs-fc5jg" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.141707 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/7d1f7d84-d053-4c95-83ac-ecbff773463d-var-lib\") pod \"ovn-controller-ovs-fc5jg\" (UID: \"7d1f7d84-d053-4c95-83ac-ecbff773463d\") " pod="openstack/ovn-controller-ovs-fc5jg" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.141751 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/7d1f7d84-d053-4c95-83ac-ecbff773463d-var-log\") pod \"ovn-controller-ovs-fc5jg\" (UID: \"7d1f7d84-d053-4c95-83ac-ecbff773463d\") " pod="openstack/ovn-controller-ovs-fc5jg" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.141813 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e8a63b50-ecd5-4993-a890-8c94bc2d5e60-var-log-ovn\") pod \"ovn-controller-cngvz\" (UID: \"e8a63b50-ecd5-4993-a890-8c94bc2d5e60\") " pod="openstack/ovn-controller-cngvz" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.141837 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e8a63b50-ecd5-4993-a890-8c94bc2d5e60-var-run-ovn\") pod \"ovn-controller-cngvz\" (UID: \"e8a63b50-ecd5-4993-a890-8c94bc2d5e60\") " pod="openstack/ovn-controller-cngvz" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.141888 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7d1f7d84-d053-4c95-83ac-ecbff773463d-var-run\") pod \"ovn-controller-ovs-fc5jg\" (UID: \"7d1f7d84-d053-4c95-83ac-ecbff773463d\") " pod="openstack/ovn-controller-ovs-fc5jg" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.141927 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e8a63b50-ecd5-4993-a890-8c94bc2d5e60-var-run\") pod \"ovn-controller-cngvz\" (UID: \"e8a63b50-ecd5-4993-a890-8c94bc2d5e60\") " pod="openstack/ovn-controller-cngvz" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.142052 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/7d1f7d84-d053-4c95-83ac-ecbff773463d-etc-ovs\") pod \"ovn-controller-ovs-fc5jg\" (UID: \"7d1f7d84-d053-4c95-83ac-ecbff773463d\") " pod="openstack/ovn-controller-ovs-fc5jg" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.144343 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e8a63b50-ecd5-4993-a890-8c94bc2d5e60-scripts\") pod \"ovn-controller-cngvz\" (UID: \"e8a63b50-ecd5-4993-a890-8c94bc2d5e60\") " pod="openstack/ovn-controller-cngvz" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.145411 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8a63b50-ecd5-4993-a890-8c94bc2d5e60-ovn-controller-tls-certs\") pod \"ovn-controller-cngvz\" (UID: \"e8a63b50-ecd5-4993-a890-8c94bc2d5e60\") " pod="openstack/ovn-controller-cngvz" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.146621 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7d1f7d84-d053-4c95-83ac-ecbff773463d-scripts\") pod \"ovn-controller-ovs-fc5jg\" (UID: \"7d1f7d84-d053-4c95-83ac-ecbff773463d\") " pod="openstack/ovn-controller-ovs-fc5jg" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.148491 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8a63b50-ecd5-4993-a890-8c94bc2d5e60-combined-ca-bundle\") pod \"ovn-controller-cngvz\" (UID: \"e8a63b50-ecd5-4993-a890-8c94bc2d5e60\") " pod="openstack/ovn-controller-cngvz" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.175949 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbdwf\" (UniqueName: \"kubernetes.io/projected/e8a63b50-ecd5-4993-a890-8c94bc2d5e60-kube-api-access-hbdwf\") pod \"ovn-controller-cngvz\" (UID: \"e8a63b50-ecd5-4993-a890-8c94bc2d5e60\") " pod="openstack/ovn-controller-cngvz" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.187993 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6vsh\" (UniqueName: \"kubernetes.io/projected/7d1f7d84-d053-4c95-83ac-ecbff773463d-kube-api-access-n6vsh\") pod \"ovn-controller-ovs-fc5jg\" (UID: \"7d1f7d84-d053-4c95-83ac-ecbff773463d\") " pod="openstack/ovn-controller-ovs-fc5jg" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.217832 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-cngvz" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.230235 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-fc5jg" Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.618794 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-kb2gr"] Sep 30 14:11:20 crc kubenswrapper[4840]: W0930 14:11:20.632021 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod67556acd_23a4_40ab_b0ba_32c267b52956.slice/crio-7912dd4657dd4126194994c2f6ea0d6c6d7dafd64d4f9e264d99894cdae4a940 WatchSource:0}: Error finding container 7912dd4657dd4126194994c2f6ea0d6c6d7dafd64d4f9e264d99894cdae4a940: Status 404 returned error can't find the container with id 7912dd4657dd4126194994c2f6ea0d6c6d7dafd64d4f9e264d99894cdae4a940 Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.874206 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.882965 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.888029 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.892604 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Sep 30 14:11:20 crc kubenswrapper[4840]: W0930 14:11:20.906472 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod41b0d67b_4ef4_4ebb_b000_b37bbdac6b34.slice/crio-cf26eb970b92a8cba1171ada9d6e56cc34dfa13a7c05df2693a58e74a47f8ecd WatchSource:0}: Error finding container cf26eb970b92a8cba1171ada9d6e56cc34dfa13a7c05df2693a58e74a47f8ecd: Status 404 returned error can't find the container with id cf26eb970b92a8cba1171ada9d6e56cc34dfa13a7c05df2693a58e74a47f8ecd Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.914736 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Sep 30 14:11:20 crc kubenswrapper[4840]: W0930 14:11:20.918764 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9eddd8f2_9e3b_43cd_ab8f_0d5c36b2dd3f.slice/crio-8c74db946473295c6a9041e39c874b404a9584591691aee9ab1557e19124d663 WatchSource:0}: Error finding container 8c74db946473295c6a9041e39c874b404a9584591691aee9ab1557e19124d663: Status 404 returned error can't find the container with id 8c74db946473295c6a9041e39c874b404a9584591691aee9ab1557e19124d663 Sep 30 14:11:20 crc kubenswrapper[4840]: W0930 14:11:20.919622 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf0207213_00d5_4b63_b041_fff191463f4a.slice/crio-81c1707399fc5db267bedac3cdd7e1e38234aebfb938b440b37b1196a41c4b51 WatchSource:0}: Error finding container 81c1707399fc5db267bedac3cdd7e1e38234aebfb938b440b37b1196a41c4b51: Status 404 returned error can't find the container with id 81c1707399fc5db267bedac3cdd7e1e38234aebfb938b440b37b1196a41c4b51 Sep 30 14:11:20 crc kubenswrapper[4840]: W0930 14:11:20.920622 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e51f892_d5b7_48b6_9a5f_0c0af8a00f21.slice/crio-46e2333b7edcfa1997bd02cd50e7f99107645217b176097071a260c0a547eeb2 WatchSource:0}: Error finding container 46e2333b7edcfa1997bd02cd50e7f99107645217b176097071a260c0a547eeb2: Status 404 returned error can't find the container with id 46e2333b7edcfa1997bd02cd50e7f99107645217b176097071a260c0a547eeb2 Sep 30 14:11:20 crc kubenswrapper[4840]: I0930 14:11:20.923604 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 30 14:11:21 crc kubenswrapper[4840]: I0930 14:11:21.142254 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-cngvz"] Sep 30 14:11:21 crc kubenswrapper[4840]: W0930 14:11:21.155496 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8a63b50_ecd5_4993_a890_8c94bc2d5e60.slice/crio-19c9852f76183caa2e94f648729829d2377cff7483d2462f89f1e94457c3b2a4 WatchSource:0}: Error finding container 19c9852f76183caa2e94f648729829d2377cff7483d2462f89f1e94457c3b2a4: Status 404 returned error can't find the container with id 19c9852f76183caa2e94f648729829d2377cff7483d2462f89f1e94457c3b2a4 Sep 30 14:11:21 crc kubenswrapper[4840]: I0930 14:11:21.225512 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Sep 30 14:11:21 crc kubenswrapper[4840]: W0930 14:11:21.227805 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2082b33d_3b6c_449a_9bdd_63d665a3d313.slice/crio-ee50c59fcf7955429a6e081167f558916e18621d7ffd6617f23fe8369c417769 WatchSource:0}: Error finding container ee50c59fcf7955429a6e081167f558916e18621d7ffd6617f23fe8369c417769: Status 404 returned error can't find the container with id ee50c59fcf7955429a6e081167f558916e18621d7ffd6617f23fe8369c417769 Sep 30 14:11:21 crc kubenswrapper[4840]: I0930 14:11:21.356532 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c15070b3-f247-4879-a7dc-618faf7e6e35","Type":"ContainerStarted","Data":"c4b33017622436b92ab437672fa03b26855f6322498a8fdc967ada50a21cd6f4"} Sep 30 14:11:21 crc kubenswrapper[4840]: I0930 14:11:21.358752 4840 generic.go:334] "Generic (PLEG): container finished" podID="8230a9ad-c799-4e92-8841-31e87f8e22fb" containerID="ad14aad5a116ee958a6557e14c9e577a2a8647fb37fcef4ecadfeba37cb29526" exitCode=0 Sep 30 14:11:21 crc kubenswrapper[4840]: I0930 14:11:21.358806 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-7f6b5" event={"ID":"8230a9ad-c799-4e92-8841-31e87f8e22fb","Type":"ContainerDied","Data":"ad14aad5a116ee958a6557e14c9e577a2a8647fb37fcef4ecadfeba37cb29526"} Sep 30 14:11:21 crc kubenswrapper[4840]: I0930 14:11:21.360539 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3e51f892-d5b7-48b6-9a5f-0c0af8a00f21","Type":"ContainerStarted","Data":"46e2333b7edcfa1997bd02cd50e7f99107645217b176097071a260c0a547eeb2"} Sep 30 14:11:21 crc kubenswrapper[4840]: I0930 14:11:21.361797 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f","Type":"ContainerStarted","Data":"8c74db946473295c6a9041e39c874b404a9584591691aee9ab1557e19124d663"} Sep 30 14:11:21 crc kubenswrapper[4840]: I0930 14:11:21.363161 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"41b0d67b-4ef4-4ebb-b000-b37bbdac6b34","Type":"ContainerStarted","Data":"cf26eb970b92a8cba1171ada9d6e56cc34dfa13a7c05df2693a58e74a47f8ecd"} Sep 30 14:11:21 crc kubenswrapper[4840]: I0930 14:11:21.364481 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b8c84ed0-e20f-445e-9548-4ce979c23476","Type":"ContainerStarted","Data":"f1a8d380d10fbaae1dc0b42b9d1439d3ba3a55ab28aff1f3e64d2fb16cbf1f05"} Sep 30 14:11:21 crc kubenswrapper[4840]: I0930 14:11:21.365829 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"2082b33d-3b6c-449a-9bdd-63d665a3d313","Type":"ContainerStarted","Data":"ee50c59fcf7955429a6e081167f558916e18621d7ffd6617f23fe8369c417769"} Sep 30 14:11:21 crc kubenswrapper[4840]: I0930 14:11:21.367318 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"f0207213-00d5-4b63-b041-fff191463f4a","Type":"ContainerStarted","Data":"81c1707399fc5db267bedac3cdd7e1e38234aebfb938b440b37b1196a41c4b51"} Sep 30 14:11:21 crc kubenswrapper[4840]: I0930 14:11:21.368908 4840 generic.go:334] "Generic (PLEG): container finished" podID="9f2e45b7-fe3e-4e53-a735-05d843a14159" containerID="74f7b95cbd5755a9bf98cf8484373a4d51bf459301fc9332cd6c18d346695043" exitCode=0 Sep 30 14:11:21 crc kubenswrapper[4840]: I0930 14:11:21.368976 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-pvrll" event={"ID":"9f2e45b7-fe3e-4e53-a735-05d843a14159","Type":"ContainerDied","Data":"74f7b95cbd5755a9bf98cf8484373a4d51bf459301fc9332cd6c18d346695043"} Sep 30 14:11:21 crc kubenswrapper[4840]: I0930 14:11:21.371765 4840 generic.go:334] "Generic (PLEG): container finished" podID="67556acd-23a4-40ab-b0ba-32c267b52956" containerID="78f0ef96141b94d78c19df5c6626ccb2afe9f70bc6326c10bec15d98df740a1d" exitCode=0 Sep 30 14:11:21 crc kubenswrapper[4840]: I0930 14:11:21.371850 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-kb2gr" event={"ID":"67556acd-23a4-40ab-b0ba-32c267b52956","Type":"ContainerDied","Data":"78f0ef96141b94d78c19df5c6626ccb2afe9f70bc6326c10bec15d98df740a1d"} Sep 30 14:11:21 crc kubenswrapper[4840]: I0930 14:11:21.372094 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-kb2gr" event={"ID":"67556acd-23a4-40ab-b0ba-32c267b52956","Type":"ContainerStarted","Data":"7912dd4657dd4126194994c2f6ea0d6c6d7dafd64d4f9e264d99894cdae4a940"} Sep 30 14:11:21 crc kubenswrapper[4840]: I0930 14:11:21.376659 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-cngvz" event={"ID":"e8a63b50-ecd5-4993-a890-8c94bc2d5e60","Type":"ContainerStarted","Data":"19c9852f76183caa2e94f648729829d2377cff7483d2462f89f1e94457c3b2a4"} Sep 30 14:11:21 crc kubenswrapper[4840]: I0930 14:11:21.386762 4840 generic.go:334] "Generic (PLEG): container finished" podID="39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2" containerID="624c899e49c9f3dcbfb78a668b8deda3475b051d38ae24fb72e56db2a59fcbc4" exitCode=0 Sep 30 14:11:21 crc kubenswrapper[4840]: I0930 14:11:21.386805 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-t9j4n" event={"ID":"39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2","Type":"ContainerDied","Data":"624c899e49c9f3dcbfb78a668b8deda3475b051d38ae24fb72e56db2a59fcbc4"} Sep 30 14:11:21 crc kubenswrapper[4840]: I0930 14:11:21.777886 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-t9j4n" Sep 30 14:11:21 crc kubenswrapper[4840]: I0930 14:11:21.784317 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-7f6b5" Sep 30 14:11:21 crc kubenswrapper[4840]: I0930 14:11:21.871975 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgllj\" (UniqueName: \"kubernetes.io/projected/8230a9ad-c799-4e92-8841-31e87f8e22fb-kube-api-access-zgllj\") pod \"8230a9ad-c799-4e92-8841-31e87f8e22fb\" (UID: \"8230a9ad-c799-4e92-8841-31e87f8e22fb\") " Sep 30 14:11:21 crc kubenswrapper[4840]: I0930 14:11:21.872151 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2-config\") pod \"39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2\" (UID: \"39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2\") " Sep 30 14:11:21 crc kubenswrapper[4840]: I0930 14:11:21.872190 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2-dns-svc\") pod \"39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2\" (UID: \"39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2\") " Sep 30 14:11:21 crc kubenswrapper[4840]: I0930 14:11:21.872244 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8230a9ad-c799-4e92-8841-31e87f8e22fb-config\") pod \"8230a9ad-c799-4e92-8841-31e87f8e22fb\" (UID: \"8230a9ad-c799-4e92-8841-31e87f8e22fb\") " Sep 30 14:11:21 crc kubenswrapper[4840]: I0930 14:11:21.872286 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2m98\" (UniqueName: \"kubernetes.io/projected/39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2-kube-api-access-n2m98\") pod \"39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2\" (UID: \"39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2\") " Sep 30 14:11:21 crc kubenswrapper[4840]: I0930 14:11:21.882510 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8230a9ad-c799-4e92-8841-31e87f8e22fb-kube-api-access-zgllj" (OuterVolumeSpecName: "kube-api-access-zgllj") pod "8230a9ad-c799-4e92-8841-31e87f8e22fb" (UID: "8230a9ad-c799-4e92-8841-31e87f8e22fb"). InnerVolumeSpecName "kube-api-access-zgllj". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:11:21 crc kubenswrapper[4840]: I0930 14:11:21.884280 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2-kube-api-access-n2m98" (OuterVolumeSpecName: "kube-api-access-n2m98") pod "39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2" (UID: "39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2"). InnerVolumeSpecName "kube-api-access-n2m98". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:11:21 crc kubenswrapper[4840]: I0930 14:11:21.899607 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2-config" (OuterVolumeSpecName: "config") pod "39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2" (UID: "39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:11:21 crc kubenswrapper[4840]: I0930 14:11:21.902988 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2" (UID: "39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:11:21 crc kubenswrapper[4840]: I0930 14:11:21.903121 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8230a9ad-c799-4e92-8841-31e87f8e22fb-config" (OuterVolumeSpecName: "config") pod "8230a9ad-c799-4e92-8841-31e87f8e22fb" (UID: "8230a9ad-c799-4e92-8841-31e87f8e22fb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:11:21 crc kubenswrapper[4840]: I0930 14:11:21.973590 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgllj\" (UniqueName: \"kubernetes.io/projected/8230a9ad-c799-4e92-8841-31e87f8e22fb-kube-api-access-zgllj\") on node \"crc\" DevicePath \"\"" Sep 30 14:11:21 crc kubenswrapper[4840]: I0930 14:11:21.973629 4840 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:11:21 crc kubenswrapper[4840]: I0930 14:11:21.973643 4840 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 14:11:21 crc kubenswrapper[4840]: I0930 14:11:21.973655 4840 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8230a9ad-c799-4e92-8841-31e87f8e22fb-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:11:21 crc kubenswrapper[4840]: I0930 14:11:21.973666 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2m98\" (UniqueName: \"kubernetes.io/projected/39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2-kube-api-access-n2m98\") on node \"crc\" DevicePath \"\"" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.001098 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-fc5jg"] Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.116227 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-2bdgj"] Sep 30 14:11:22 crc kubenswrapper[4840]: E0930 14:11:22.116920 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8230a9ad-c799-4e92-8841-31e87f8e22fb" containerName="init" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.116945 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="8230a9ad-c799-4e92-8841-31e87f8e22fb" containerName="init" Sep 30 14:11:22 crc kubenswrapper[4840]: E0930 14:11:22.116988 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2" containerName="init" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.116998 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2" containerName="init" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.117198 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2" containerName="init" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.117291 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="8230a9ad-c799-4e92-8841-31e87f8e22fb" containerName="init" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.121945 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-2bdgj" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.128957 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.153251 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-2bdgj"] Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.283763 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a789c20a-6871-4866-bfa2-c69efd876afb-combined-ca-bundle\") pod \"ovn-controller-metrics-2bdgj\" (UID: \"a789c20a-6871-4866-bfa2-c69efd876afb\") " pod="openstack/ovn-controller-metrics-2bdgj" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.283844 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a789c20a-6871-4866-bfa2-c69efd876afb-ovs-rundir\") pod \"ovn-controller-metrics-2bdgj\" (UID: \"a789c20a-6871-4866-bfa2-c69efd876afb\") " pod="openstack/ovn-controller-metrics-2bdgj" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.283906 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a789c20a-6871-4866-bfa2-c69efd876afb-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-2bdgj\" (UID: \"a789c20a-6871-4866-bfa2-c69efd876afb\") " pod="openstack/ovn-controller-metrics-2bdgj" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.284044 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a789c20a-6871-4866-bfa2-c69efd876afb-config\") pod \"ovn-controller-metrics-2bdgj\" (UID: \"a789c20a-6871-4866-bfa2-c69efd876afb\") " pod="openstack/ovn-controller-metrics-2bdgj" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.284141 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmnvs\" (UniqueName: \"kubernetes.io/projected/a789c20a-6871-4866-bfa2-c69efd876afb-kube-api-access-wmnvs\") pod \"ovn-controller-metrics-2bdgj\" (UID: \"a789c20a-6871-4866-bfa2-c69efd876afb\") " pod="openstack/ovn-controller-metrics-2bdgj" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.284306 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a789c20a-6871-4866-bfa2-c69efd876afb-ovn-rundir\") pod \"ovn-controller-metrics-2bdgj\" (UID: \"a789c20a-6871-4866-bfa2-c69efd876afb\") " pod="openstack/ovn-controller-metrics-2bdgj" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.286971 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-kb2gr"] Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.325750 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-wplqh"] Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.326996 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-wplqh" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.330451 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.349291 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-wplqh"] Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.386084 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a789c20a-6871-4866-bfa2-c69efd876afb-ovn-rundir\") pod \"ovn-controller-metrics-2bdgj\" (UID: \"a789c20a-6871-4866-bfa2-c69efd876afb\") " pod="openstack/ovn-controller-metrics-2bdgj" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.386159 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a789c20a-6871-4866-bfa2-c69efd876afb-combined-ca-bundle\") pod \"ovn-controller-metrics-2bdgj\" (UID: \"a789c20a-6871-4866-bfa2-c69efd876afb\") " pod="openstack/ovn-controller-metrics-2bdgj" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.386206 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a789c20a-6871-4866-bfa2-c69efd876afb-ovs-rundir\") pod \"ovn-controller-metrics-2bdgj\" (UID: \"a789c20a-6871-4866-bfa2-c69efd876afb\") " pod="openstack/ovn-controller-metrics-2bdgj" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.386250 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a789c20a-6871-4866-bfa2-c69efd876afb-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-2bdgj\" (UID: \"a789c20a-6871-4866-bfa2-c69efd876afb\") " pod="openstack/ovn-controller-metrics-2bdgj" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.386278 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a789c20a-6871-4866-bfa2-c69efd876afb-config\") pod \"ovn-controller-metrics-2bdgj\" (UID: \"a789c20a-6871-4866-bfa2-c69efd876afb\") " pod="openstack/ovn-controller-metrics-2bdgj" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.386294 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmnvs\" (UniqueName: \"kubernetes.io/projected/a789c20a-6871-4866-bfa2-c69efd876afb-kube-api-access-wmnvs\") pod \"ovn-controller-metrics-2bdgj\" (UID: \"a789c20a-6871-4866-bfa2-c69efd876afb\") " pod="openstack/ovn-controller-metrics-2bdgj" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.386433 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a789c20a-6871-4866-bfa2-c69efd876afb-ovn-rundir\") pod \"ovn-controller-metrics-2bdgj\" (UID: \"a789c20a-6871-4866-bfa2-c69efd876afb\") " pod="openstack/ovn-controller-metrics-2bdgj" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.387149 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a789c20a-6871-4866-bfa2-c69efd876afb-ovs-rundir\") pod \"ovn-controller-metrics-2bdgj\" (UID: \"a789c20a-6871-4866-bfa2-c69efd876afb\") " pod="openstack/ovn-controller-metrics-2bdgj" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.388216 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a789c20a-6871-4866-bfa2-c69efd876afb-config\") pod \"ovn-controller-metrics-2bdgj\" (UID: \"a789c20a-6871-4866-bfa2-c69efd876afb\") " pod="openstack/ovn-controller-metrics-2bdgj" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.409595 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a789c20a-6871-4866-bfa2-c69efd876afb-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-2bdgj\" (UID: \"a789c20a-6871-4866-bfa2-c69efd876afb\") " pod="openstack/ovn-controller-metrics-2bdgj" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.415544 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a789c20a-6871-4866-bfa2-c69efd876afb-combined-ca-bundle\") pod \"ovn-controller-metrics-2bdgj\" (UID: \"a789c20a-6871-4866-bfa2-c69efd876afb\") " pod="openstack/ovn-controller-metrics-2bdgj" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.419614 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-7f6b5" event={"ID":"8230a9ad-c799-4e92-8841-31e87f8e22fb","Type":"ContainerDied","Data":"6dda08756fd4db73c84728907e67c65474ded83750972327a8bd47ea80db0ecc"} Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.419678 4840 scope.go:117] "RemoveContainer" containerID="ad14aad5a116ee958a6557e14c9e577a2a8647fb37fcef4ecadfeba37cb29526" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.419824 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-7f6b5" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.421193 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmnvs\" (UniqueName: \"kubernetes.io/projected/a789c20a-6871-4866-bfa2-c69efd876afb-kube-api-access-wmnvs\") pod \"ovn-controller-metrics-2bdgj\" (UID: \"a789c20a-6871-4866-bfa2-c69efd876afb\") " pod="openstack/ovn-controller-metrics-2bdgj" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.432382 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-pvrll" event={"ID":"9f2e45b7-fe3e-4e53-a735-05d843a14159","Type":"ContainerStarted","Data":"b6817e74a280d4642bbb241ea913659cb63427fbe679a334fe8dfdf5f1a1188d"} Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.433647 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5ccc8479f9-pvrll" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.456277 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-kb2gr" event={"ID":"67556acd-23a4-40ab-b0ba-32c267b52956","Type":"ContainerStarted","Data":"4a56f3accbb2dfa7e7a6c5cea2bc5f2955f7fe246c708cf626f053b7044e1b73"} Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.457652 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-kb2gr" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.459310 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-2bdgj" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.463512 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5ccc8479f9-pvrll" podStartSLOduration=5.127663849 podStartE2EDuration="13.463495244s" podCreationTimestamp="2025-09-30 14:11:09 +0000 UTC" firstStartedPulling="2025-09-30 14:11:12.00090569 +0000 UTC m=+900.629992113" lastFinishedPulling="2025-09-30 14:11:20.336737085 +0000 UTC m=+908.965823508" observedRunningTime="2025-09-30 14:11:22.458247397 +0000 UTC m=+911.087333820" watchObservedRunningTime="2025-09-30 14:11:22.463495244 +0000 UTC m=+911.092581667" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.479665 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-kb2gr" podStartSLOduration=13.479638934 podStartE2EDuration="13.479638934s" podCreationTimestamp="2025-09-30 14:11:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:11:22.476636596 +0000 UTC m=+911.105723019" watchObservedRunningTime="2025-09-30 14:11:22.479638934 +0000 UTC m=+911.108725377" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.481160 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-t9j4n" event={"ID":"39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2","Type":"ContainerDied","Data":"c623c300c431eb9ba4a2f0f8be3ff1e401ab2d8a40abb637a802d520e08117d1"} Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.481213 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-t9j4n" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.487437 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5cbf513e-7178-4a46-bc58-c3bfcfa728ff-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-wplqh\" (UID: \"5cbf513e-7178-4a46-bc58-c3bfcfa728ff\") " pod="openstack/dnsmasq-dns-7fd796d7df-wplqh" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.487492 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5cbf513e-7178-4a46-bc58-c3bfcfa728ff-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-wplqh\" (UID: \"5cbf513e-7178-4a46-bc58-c3bfcfa728ff\") " pod="openstack/dnsmasq-dns-7fd796d7df-wplqh" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.487531 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksc66\" (UniqueName: \"kubernetes.io/projected/5cbf513e-7178-4a46-bc58-c3bfcfa728ff-kube-api-access-ksc66\") pod \"dnsmasq-dns-7fd796d7df-wplqh\" (UID: \"5cbf513e-7178-4a46-bc58-c3bfcfa728ff\") " pod="openstack/dnsmasq-dns-7fd796d7df-wplqh" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.487583 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cbf513e-7178-4a46-bc58-c3bfcfa728ff-config\") pod \"dnsmasq-dns-7fd796d7df-wplqh\" (UID: \"5cbf513e-7178-4a46-bc58-c3bfcfa728ff\") " pod="openstack/dnsmasq-dns-7fd796d7df-wplqh" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.521546 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-7f6b5"] Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.542502 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-7f6b5"] Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.567405 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-t9j4n"] Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.572201 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-t9j4n"] Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.588567 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5cbf513e-7178-4a46-bc58-c3bfcfa728ff-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-wplqh\" (UID: \"5cbf513e-7178-4a46-bc58-c3bfcfa728ff\") " pod="openstack/dnsmasq-dns-7fd796d7df-wplqh" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.588613 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5cbf513e-7178-4a46-bc58-c3bfcfa728ff-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-wplqh\" (UID: \"5cbf513e-7178-4a46-bc58-c3bfcfa728ff\") " pod="openstack/dnsmasq-dns-7fd796d7df-wplqh" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.588652 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksc66\" (UniqueName: \"kubernetes.io/projected/5cbf513e-7178-4a46-bc58-c3bfcfa728ff-kube-api-access-ksc66\") pod \"dnsmasq-dns-7fd796d7df-wplqh\" (UID: \"5cbf513e-7178-4a46-bc58-c3bfcfa728ff\") " pod="openstack/dnsmasq-dns-7fd796d7df-wplqh" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.588695 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cbf513e-7178-4a46-bc58-c3bfcfa728ff-config\") pod \"dnsmasq-dns-7fd796d7df-wplqh\" (UID: \"5cbf513e-7178-4a46-bc58-c3bfcfa728ff\") " pod="openstack/dnsmasq-dns-7fd796d7df-wplqh" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.589590 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cbf513e-7178-4a46-bc58-c3bfcfa728ff-config\") pod \"dnsmasq-dns-7fd796d7df-wplqh\" (UID: \"5cbf513e-7178-4a46-bc58-c3bfcfa728ff\") " pod="openstack/dnsmasq-dns-7fd796d7df-wplqh" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.589795 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5cbf513e-7178-4a46-bc58-c3bfcfa728ff-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-wplqh\" (UID: \"5cbf513e-7178-4a46-bc58-c3bfcfa728ff\") " pod="openstack/dnsmasq-dns-7fd796d7df-wplqh" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.590492 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5cbf513e-7178-4a46-bc58-c3bfcfa728ff-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-wplqh\" (UID: \"5cbf513e-7178-4a46-bc58-c3bfcfa728ff\") " pod="openstack/dnsmasq-dns-7fd796d7df-wplqh" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.616729 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksc66\" (UniqueName: \"kubernetes.io/projected/5cbf513e-7178-4a46-bc58-c3bfcfa728ff-kube-api-access-ksc66\") pod \"dnsmasq-dns-7fd796d7df-wplqh\" (UID: \"5cbf513e-7178-4a46-bc58-c3bfcfa728ff\") " pod="openstack/dnsmasq-dns-7fd796d7df-wplqh" Sep 30 14:11:22 crc kubenswrapper[4840]: I0930 14:11:22.654526 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-wplqh" Sep 30 14:11:23 crc kubenswrapper[4840]: I0930 14:11:23.043165 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Sep 30 14:11:23 crc kubenswrapper[4840]: I0930 14:11:23.045466 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Sep 30 14:11:23 crc kubenswrapper[4840]: I0930 14:11:23.049342 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Sep 30 14:11:23 crc kubenswrapper[4840]: I0930 14:11:23.049809 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-jh9dr" Sep 30 14:11:23 crc kubenswrapper[4840]: I0930 14:11:23.050031 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Sep 30 14:11:23 crc kubenswrapper[4840]: I0930 14:11:23.050141 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Sep 30 14:11:23 crc kubenswrapper[4840]: I0930 14:11:23.056778 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Sep 30 14:11:23 crc kubenswrapper[4840]: I0930 14:11:23.215572 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/93d37b01-5288-4d31-84db-3fd7e7840c3b-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"93d37b01-5288-4d31-84db-3fd7e7840c3b\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:11:23 crc kubenswrapper[4840]: I0930 14:11:23.215945 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/93d37b01-5288-4d31-84db-3fd7e7840c3b-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"93d37b01-5288-4d31-84db-3fd7e7840c3b\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:11:23 crc kubenswrapper[4840]: I0930 14:11:23.215986 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"93d37b01-5288-4d31-84db-3fd7e7840c3b\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:11:23 crc kubenswrapper[4840]: I0930 14:11:23.216032 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93d37b01-5288-4d31-84db-3fd7e7840c3b-config\") pod \"ovsdbserver-sb-0\" (UID: \"93d37b01-5288-4d31-84db-3fd7e7840c3b\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:11:23 crc kubenswrapper[4840]: I0930 14:11:23.216047 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/93d37b01-5288-4d31-84db-3fd7e7840c3b-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"93d37b01-5288-4d31-84db-3fd7e7840c3b\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:11:23 crc kubenswrapper[4840]: I0930 14:11:23.216068 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93d37b01-5288-4d31-84db-3fd7e7840c3b-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"93d37b01-5288-4d31-84db-3fd7e7840c3b\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:11:23 crc kubenswrapper[4840]: I0930 14:11:23.216108 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zcsr\" (UniqueName: \"kubernetes.io/projected/93d37b01-5288-4d31-84db-3fd7e7840c3b-kube-api-access-9zcsr\") pod \"ovsdbserver-sb-0\" (UID: \"93d37b01-5288-4d31-84db-3fd7e7840c3b\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:11:23 crc kubenswrapper[4840]: I0930 14:11:23.216146 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/93d37b01-5288-4d31-84db-3fd7e7840c3b-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"93d37b01-5288-4d31-84db-3fd7e7840c3b\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:11:23 crc kubenswrapper[4840]: I0930 14:11:23.318617 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/93d37b01-5288-4d31-84db-3fd7e7840c3b-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"93d37b01-5288-4d31-84db-3fd7e7840c3b\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:11:23 crc kubenswrapper[4840]: I0930 14:11:23.318705 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"93d37b01-5288-4d31-84db-3fd7e7840c3b\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:11:23 crc kubenswrapper[4840]: I0930 14:11:23.318782 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/93d37b01-5288-4d31-84db-3fd7e7840c3b-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"93d37b01-5288-4d31-84db-3fd7e7840c3b\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:11:23 crc kubenswrapper[4840]: I0930 14:11:23.318801 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93d37b01-5288-4d31-84db-3fd7e7840c3b-config\") pod \"ovsdbserver-sb-0\" (UID: \"93d37b01-5288-4d31-84db-3fd7e7840c3b\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:11:23 crc kubenswrapper[4840]: I0930 14:11:23.318840 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93d37b01-5288-4d31-84db-3fd7e7840c3b-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"93d37b01-5288-4d31-84db-3fd7e7840c3b\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:11:23 crc kubenswrapper[4840]: I0930 14:11:23.318887 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zcsr\" (UniqueName: \"kubernetes.io/projected/93d37b01-5288-4d31-84db-3fd7e7840c3b-kube-api-access-9zcsr\") pod \"ovsdbserver-sb-0\" (UID: \"93d37b01-5288-4d31-84db-3fd7e7840c3b\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:11:23 crc kubenswrapper[4840]: I0930 14:11:23.318945 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/93d37b01-5288-4d31-84db-3fd7e7840c3b-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"93d37b01-5288-4d31-84db-3fd7e7840c3b\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:11:23 crc kubenswrapper[4840]: I0930 14:11:23.319028 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/93d37b01-5288-4d31-84db-3fd7e7840c3b-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"93d37b01-5288-4d31-84db-3fd7e7840c3b\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:11:23 crc kubenswrapper[4840]: I0930 14:11:23.320095 4840 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"93d37b01-5288-4d31-84db-3fd7e7840c3b\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/ovsdbserver-sb-0" Sep 30 14:11:23 crc kubenswrapper[4840]: I0930 14:11:23.321143 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/93d37b01-5288-4d31-84db-3fd7e7840c3b-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"93d37b01-5288-4d31-84db-3fd7e7840c3b\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:11:23 crc kubenswrapper[4840]: I0930 14:11:23.321670 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93d37b01-5288-4d31-84db-3fd7e7840c3b-config\") pod \"ovsdbserver-sb-0\" (UID: \"93d37b01-5288-4d31-84db-3fd7e7840c3b\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:11:23 crc kubenswrapper[4840]: I0930 14:11:23.322427 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/93d37b01-5288-4d31-84db-3fd7e7840c3b-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"93d37b01-5288-4d31-84db-3fd7e7840c3b\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:11:23 crc kubenswrapper[4840]: I0930 14:11:23.325515 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/93d37b01-5288-4d31-84db-3fd7e7840c3b-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"93d37b01-5288-4d31-84db-3fd7e7840c3b\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:11:23 crc kubenswrapper[4840]: I0930 14:11:23.326506 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/93d37b01-5288-4d31-84db-3fd7e7840c3b-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"93d37b01-5288-4d31-84db-3fd7e7840c3b\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:11:23 crc kubenswrapper[4840]: I0930 14:11:23.331308 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93d37b01-5288-4d31-84db-3fd7e7840c3b-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"93d37b01-5288-4d31-84db-3fd7e7840c3b\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:11:23 crc kubenswrapper[4840]: I0930 14:11:23.341081 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zcsr\" (UniqueName: \"kubernetes.io/projected/93d37b01-5288-4d31-84db-3fd7e7840c3b-kube-api-access-9zcsr\") pod \"ovsdbserver-sb-0\" (UID: \"93d37b01-5288-4d31-84db-3fd7e7840c3b\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:11:23 crc kubenswrapper[4840]: I0930 14:11:23.342278 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"93d37b01-5288-4d31-84db-3fd7e7840c3b\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:11:23 crc kubenswrapper[4840]: I0930 14:11:23.373055 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Sep 30 14:11:23 crc kubenswrapper[4840]: I0930 14:11:23.488150 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-kb2gr" podUID="67556acd-23a4-40ab-b0ba-32c267b52956" containerName="dnsmasq-dns" containerID="cri-o://4a56f3accbb2dfa7e7a6c5cea2bc5f2955f7fe246c708cf626f053b7044e1b73" gracePeriod=10 Sep 30 14:11:24 crc kubenswrapper[4840]: I0930 14:11:24.128275 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2" path="/var/lib/kubelet/pods/39ff3a7e-fad5-4ca6-94d1-68ba2d2fdba2/volumes" Sep 30 14:11:24 crc kubenswrapper[4840]: I0930 14:11:24.128802 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8230a9ad-c799-4e92-8841-31e87f8e22fb" path="/var/lib/kubelet/pods/8230a9ad-c799-4e92-8841-31e87f8e22fb/volumes" Sep 30 14:11:24 crc kubenswrapper[4840]: W0930 14:11:24.928191 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d1f7d84_d053_4c95_83ac_ecbff773463d.slice/crio-d01202201538fe1528a97954d488bf3a4e841debda0da013b3013bd9267099d8 WatchSource:0}: Error finding container d01202201538fe1528a97954d488bf3a4e841debda0da013b3013bd9267099d8: Status 404 returned error can't find the container with id d01202201538fe1528a97954d488bf3a4e841debda0da013b3013bd9267099d8 Sep 30 14:11:25 crc kubenswrapper[4840]: I0930 14:11:25.501957 4840 generic.go:334] "Generic (PLEG): container finished" podID="67556acd-23a4-40ab-b0ba-32c267b52956" containerID="4a56f3accbb2dfa7e7a6c5cea2bc5f2955f7fe246c708cf626f053b7044e1b73" exitCode=0 Sep 30 14:11:25 crc kubenswrapper[4840]: I0930 14:11:25.502043 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-kb2gr" event={"ID":"67556acd-23a4-40ab-b0ba-32c267b52956","Type":"ContainerDied","Data":"4a56f3accbb2dfa7e7a6c5cea2bc5f2955f7fe246c708cf626f053b7044e1b73"} Sep 30 14:11:25 crc kubenswrapper[4840]: I0930 14:11:25.506589 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-fc5jg" event={"ID":"7d1f7d84-d053-4c95-83ac-ecbff773463d","Type":"ContainerStarted","Data":"d01202201538fe1528a97954d488bf3a4e841debda0da013b3013bd9267099d8"} Sep 30 14:11:25 crc kubenswrapper[4840]: I0930 14:11:25.905260 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-kb2gr" Sep 30 14:11:26 crc kubenswrapper[4840]: I0930 14:11:26.061415 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zg9qm\" (UniqueName: \"kubernetes.io/projected/67556acd-23a4-40ab-b0ba-32c267b52956-kube-api-access-zg9qm\") pod \"67556acd-23a4-40ab-b0ba-32c267b52956\" (UID: \"67556acd-23a4-40ab-b0ba-32c267b52956\") " Sep 30 14:11:26 crc kubenswrapper[4840]: I0930 14:11:26.061511 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67556acd-23a4-40ab-b0ba-32c267b52956-dns-svc\") pod \"67556acd-23a4-40ab-b0ba-32c267b52956\" (UID: \"67556acd-23a4-40ab-b0ba-32c267b52956\") " Sep 30 14:11:26 crc kubenswrapper[4840]: I0930 14:11:26.061542 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67556acd-23a4-40ab-b0ba-32c267b52956-config\") pod \"67556acd-23a4-40ab-b0ba-32c267b52956\" (UID: \"67556acd-23a4-40ab-b0ba-32c267b52956\") " Sep 30 14:11:26 crc kubenswrapper[4840]: I0930 14:11:26.067121 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67556acd-23a4-40ab-b0ba-32c267b52956-kube-api-access-zg9qm" (OuterVolumeSpecName: "kube-api-access-zg9qm") pod "67556acd-23a4-40ab-b0ba-32c267b52956" (UID: "67556acd-23a4-40ab-b0ba-32c267b52956"). InnerVolumeSpecName "kube-api-access-zg9qm". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:11:26 crc kubenswrapper[4840]: I0930 14:11:26.104458 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67556acd-23a4-40ab-b0ba-32c267b52956-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "67556acd-23a4-40ab-b0ba-32c267b52956" (UID: "67556acd-23a4-40ab-b0ba-32c267b52956"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:11:26 crc kubenswrapper[4840]: I0930 14:11:26.108037 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67556acd-23a4-40ab-b0ba-32c267b52956-config" (OuterVolumeSpecName: "config") pod "67556acd-23a4-40ab-b0ba-32c267b52956" (UID: "67556acd-23a4-40ab-b0ba-32c267b52956"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:11:26 crc kubenswrapper[4840]: I0930 14:11:26.162942 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zg9qm\" (UniqueName: \"kubernetes.io/projected/67556acd-23a4-40ab-b0ba-32c267b52956-kube-api-access-zg9qm\") on node \"crc\" DevicePath \"\"" Sep 30 14:11:26 crc kubenswrapper[4840]: I0930 14:11:26.162972 4840 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67556acd-23a4-40ab-b0ba-32c267b52956-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 14:11:26 crc kubenswrapper[4840]: I0930 14:11:26.162982 4840 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67556acd-23a4-40ab-b0ba-32c267b52956-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:11:26 crc kubenswrapper[4840]: I0930 14:11:26.517970 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-kb2gr" event={"ID":"67556acd-23a4-40ab-b0ba-32c267b52956","Type":"ContainerDied","Data":"7912dd4657dd4126194994c2f6ea0d6c6d7dafd64d4f9e264d99894cdae4a940"} Sep 30 14:11:26 crc kubenswrapper[4840]: I0930 14:11:26.518064 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-kb2gr" Sep 30 14:11:26 crc kubenswrapper[4840]: I0930 14:11:26.540671 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-kb2gr"] Sep 30 14:11:26 crc kubenswrapper[4840]: I0930 14:11:26.546761 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-kb2gr"] Sep 30 14:11:26 crc kubenswrapper[4840]: I0930 14:11:26.928510 4840 scope.go:117] "RemoveContainer" containerID="624c899e49c9f3dcbfb78a668b8deda3475b051d38ae24fb72e56db2a59fcbc4" Sep 30 14:11:28 crc kubenswrapper[4840]: I0930 14:11:28.126242 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67556acd-23a4-40ab-b0ba-32c267b52956" path="/var/lib/kubelet/pods/67556acd-23a4-40ab-b0ba-32c267b52956/volumes" Sep 30 14:11:29 crc kubenswrapper[4840]: I0930 14:11:29.521994 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5ccc8479f9-pvrll" Sep 30 14:11:30 crc kubenswrapper[4840]: I0930 14:11:30.104997 4840 scope.go:117] "RemoveContainer" containerID="4a56f3accbb2dfa7e7a6c5cea2bc5f2955f7fe246c708cf626f053b7044e1b73" Sep 30 14:11:30 crc kubenswrapper[4840]: I0930 14:11:30.151048 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-wplqh"] Sep 30 14:11:30 crc kubenswrapper[4840]: W0930 14:11:30.404257 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5cbf513e_7178_4a46_bc58_c3bfcfa728ff.slice/crio-ebd48ad79841e44296b4299bccc3357e0379ab3822f7c6da68121734a42c4abd WatchSource:0}: Error finding container ebd48ad79841e44296b4299bccc3357e0379ab3822f7c6da68121734a42c4abd: Status 404 returned error can't find the container with id ebd48ad79841e44296b4299bccc3357e0379ab3822f7c6da68121734a42c4abd Sep 30 14:11:30 crc kubenswrapper[4840]: I0930 14:11:30.514783 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-2bdgj"] Sep 30 14:11:30 crc kubenswrapper[4840]: I0930 14:11:30.521564 4840 scope.go:117] "RemoveContainer" containerID="78f0ef96141b94d78c19df5c6626ccb2afe9f70bc6326c10bec15d98df740a1d" Sep 30 14:11:30 crc kubenswrapper[4840]: I0930 14:11:30.556211 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-wplqh" event={"ID":"5cbf513e-7178-4a46-bc58-c3bfcfa728ff","Type":"ContainerStarted","Data":"ebd48ad79841e44296b4299bccc3357e0379ab3822f7c6da68121734a42c4abd"} Sep 30 14:11:30 crc kubenswrapper[4840]: I0930 14:11:30.560633 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-2bdgj" event={"ID":"a789c20a-6871-4866-bfa2-c69efd876afb","Type":"ContainerStarted","Data":"210958547c620a0c10a1ca3f5ef2b4ed28784b91d0817b2b7b447a460ecfeeb1"} Sep 30 14:11:30 crc kubenswrapper[4840]: I0930 14:11:30.857183 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Sep 30 14:11:30 crc kubenswrapper[4840]: W0930 14:11:30.912890 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93d37b01_5288_4d31_84db_3fd7e7840c3b.slice/crio-61ccd60b892c1fc1b0548985ff6b84677daf87c8e2624d2c5320cc319dcb398a WatchSource:0}: Error finding container 61ccd60b892c1fc1b0548985ff6b84677daf87c8e2624d2c5320cc319dcb398a: Status 404 returned error can't find the container with id 61ccd60b892c1fc1b0548985ff6b84677daf87c8e2624d2c5320cc319dcb398a Sep 30 14:11:31 crc kubenswrapper[4840]: I0930 14:11:31.582589 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b8c84ed0-e20f-445e-9548-4ce979c23476","Type":"ContainerStarted","Data":"d2cdd432cff6329b6a00d90cea6d034e25aced440092be6ed044b88f6112f277"} Sep 30 14:11:31 crc kubenswrapper[4840]: I0930 14:11:31.586659 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"2082b33d-3b6c-449a-9bdd-63d665a3d313","Type":"ContainerStarted","Data":"805b7d3cf49d4f275fd3e63b3e8468e63b1da9634478dc84b2eff3a581f27dca"} Sep 30 14:11:31 crc kubenswrapper[4840]: I0930 14:11:31.594201 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"f0207213-00d5-4b63-b041-fff191463f4a","Type":"ContainerStarted","Data":"531da2c69c85155c80a112857c5c05f82f68dcdbd38265d6170fd1cbfe9726b6"} Sep 30 14:11:31 crc kubenswrapper[4840]: I0930 14:11:31.594323 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Sep 30 14:11:31 crc kubenswrapper[4840]: I0930 14:11:31.597408 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"93d37b01-5288-4d31-84db-3fd7e7840c3b","Type":"ContainerStarted","Data":"61ccd60b892c1fc1b0548985ff6b84677daf87c8e2624d2c5320cc319dcb398a"} Sep 30 14:11:31 crc kubenswrapper[4840]: I0930 14:11:31.609223 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f","Type":"ContainerStarted","Data":"05ac69d0f33558ccad10e1fe8b72e19be47e55f4559138a238367ed19f65466d"} Sep 30 14:11:31 crc kubenswrapper[4840]: I0930 14:11:31.611951 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"41b0d67b-4ef4-4ebb-b000-b37bbdac6b34","Type":"ContainerStarted","Data":"b13172bf666fdd4d12af9ec2bb0f60e000d1d9ae612fc8995e6b5757136cfee3"} Sep 30 14:11:31 crc kubenswrapper[4840]: I0930 14:11:31.639439 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=9.940429716 podStartE2EDuration="18.639423522s" podCreationTimestamp="2025-09-30 14:11:13 +0000 UTC" firstStartedPulling="2025-09-30 14:11:20.925682529 +0000 UTC m=+909.554768952" lastFinishedPulling="2025-09-30 14:11:29.624676325 +0000 UTC m=+918.253762758" observedRunningTime="2025-09-30 14:11:31.634300949 +0000 UTC m=+920.263387372" watchObservedRunningTime="2025-09-30 14:11:31.639423522 +0000 UTC m=+920.268509945" Sep 30 14:11:32 crc kubenswrapper[4840]: I0930 14:11:32.624852 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-cngvz" event={"ID":"e8a63b50-ecd5-4993-a890-8c94bc2d5e60","Type":"ContainerStarted","Data":"45b92ade658266d3428479f8a790f391451eced6270157f6b810e539991c974a"} Sep 30 14:11:32 crc kubenswrapper[4840]: I0930 14:11:32.625511 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-cngvz" Sep 30 14:11:32 crc kubenswrapper[4840]: I0930 14:11:32.629822 4840 generic.go:334] "Generic (PLEG): container finished" podID="7d1f7d84-d053-4c95-83ac-ecbff773463d" containerID="dedf159217a180d2feb98eb3b65bf4629b28ad6af57536f32f574802054ea783" exitCode=0 Sep 30 14:11:32 crc kubenswrapper[4840]: I0930 14:11:32.629922 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-fc5jg" event={"ID":"7d1f7d84-d053-4c95-83ac-ecbff773463d","Type":"ContainerDied","Data":"dedf159217a180d2feb98eb3b65bf4629b28ad6af57536f32f574802054ea783"} Sep 30 14:11:32 crc kubenswrapper[4840]: I0930 14:11:32.633212 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c15070b3-f247-4879-a7dc-618faf7e6e35","Type":"ContainerStarted","Data":"10b078a915c98d66021e33eaf12b5a5033f297335c7c69779a5f20a19252eb1c"} Sep 30 14:11:32 crc kubenswrapper[4840]: I0930 14:11:32.643931 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"93d37b01-5288-4d31-84db-3fd7e7840c3b","Type":"ContainerStarted","Data":"5e2eba0281ecd43911bd8cc53e614c109a039f7c1ecd6a0ce54db930e88a2dea"} Sep 30 14:11:32 crc kubenswrapper[4840]: I0930 14:11:32.645736 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-cngvz" podStartSLOduration=3.900164198 podStartE2EDuration="13.645712457s" podCreationTimestamp="2025-09-30 14:11:19 +0000 UTC" firstStartedPulling="2025-09-30 14:11:21.157365133 +0000 UTC m=+909.786451556" lastFinishedPulling="2025-09-30 14:11:30.902913392 +0000 UTC m=+919.531999815" observedRunningTime="2025-09-30 14:11:32.643169981 +0000 UTC m=+921.272256414" watchObservedRunningTime="2025-09-30 14:11:32.645712457 +0000 UTC m=+921.274798880" Sep 30 14:11:32 crc kubenswrapper[4840]: I0930 14:11:32.670075 4840 generic.go:334] "Generic (PLEG): container finished" podID="5cbf513e-7178-4a46-bc58-c3bfcfa728ff" containerID="4f111f857791503b05a97de781e9f61a7de7f7f6f0420efd87a6878c46513080" exitCode=0 Sep 30 14:11:32 crc kubenswrapper[4840]: I0930 14:11:32.670194 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-wplqh" event={"ID":"5cbf513e-7178-4a46-bc58-c3bfcfa728ff","Type":"ContainerDied","Data":"4f111f857791503b05a97de781e9f61a7de7f7f6f0420efd87a6878c46513080"} Sep 30 14:11:32 crc kubenswrapper[4840]: I0930 14:11:32.676890 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3e51f892-d5b7-48b6-9a5f-0c0af8a00f21","Type":"ContainerStarted","Data":"a8c84f922f39612d031013b9885fdcb52dcc7faafb7a80671ecfc2c1353cc4e6"} Sep 30 14:11:32 crc kubenswrapper[4840]: I0930 14:11:32.677136 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Sep 30 14:11:32 crc kubenswrapper[4840]: I0930 14:11:32.722927 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=7.355119666 podStartE2EDuration="17.722907414s" podCreationTimestamp="2025-09-30 14:11:15 +0000 UTC" firstStartedPulling="2025-09-30 14:11:20.924376385 +0000 UTC m=+909.553462808" lastFinishedPulling="2025-09-30 14:11:31.292164133 +0000 UTC m=+919.921250556" observedRunningTime="2025-09-30 14:11:32.720104861 +0000 UTC m=+921.349191284" watchObservedRunningTime="2025-09-30 14:11:32.722907414 +0000 UTC m=+921.351993847" Sep 30 14:11:34 crc kubenswrapper[4840]: I0930 14:11:34.694637 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-2bdgj" event={"ID":"a789c20a-6871-4866-bfa2-c69efd876afb","Type":"ContainerStarted","Data":"0ddf629864b84b62eaeab8bb6f943a8ea92cbcb87176e3436ed1f1ed218656dd"} Sep 30 14:11:34 crc kubenswrapper[4840]: I0930 14:11:34.698249 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-fc5jg" event={"ID":"7d1f7d84-d053-4c95-83ac-ecbff773463d","Type":"ContainerStarted","Data":"3520d6268b3a8324b21a4016202e8d408f25adcb82be93ef9f22116e0f21687f"} Sep 30 14:11:34 crc kubenswrapper[4840]: I0930 14:11:34.698299 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-fc5jg" event={"ID":"7d1f7d84-d053-4c95-83ac-ecbff773463d","Type":"ContainerStarted","Data":"c3bb8853430089fd51704d88e12fad25c7cac054d717c04b92da5928207fb604"} Sep 30 14:11:34 crc kubenswrapper[4840]: I0930 14:11:34.700579 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"2082b33d-3b6c-449a-9bdd-63d665a3d313","Type":"ContainerStarted","Data":"d582d6688bc430c1302286d59daeedeb915ff91e08cf9340f3fcc7ae4a629f64"} Sep 30 14:11:34 crc kubenswrapper[4840]: I0930 14:11:34.702519 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"93d37b01-5288-4d31-84db-3fd7e7840c3b","Type":"ContainerStarted","Data":"41b9aa143a3c69eb5cfdc38fbc9af2843a34b78c3c5f77a82fa333d77d74e138"} Sep 30 14:11:34 crc kubenswrapper[4840]: I0930 14:11:34.704985 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-wplqh" event={"ID":"5cbf513e-7178-4a46-bc58-c3bfcfa728ff","Type":"ContainerStarted","Data":"86e99e9ad660a3084b5ba24a7d04f4907f4359ad9abf610b2b14ab823d5d7a02"} Sep 30 14:11:34 crc kubenswrapper[4840]: I0930 14:11:34.705122 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7fd796d7df-wplqh" Sep 30 14:11:34 crc kubenswrapper[4840]: I0930 14:11:34.715684 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-2bdgj" podStartSLOduration=9.298451627 podStartE2EDuration="12.715662329s" podCreationTimestamp="2025-09-30 14:11:22 +0000 UTC" firstStartedPulling="2025-09-30 14:11:30.530617972 +0000 UTC m=+919.159704385" lastFinishedPulling="2025-09-30 14:11:33.947828664 +0000 UTC m=+922.576915087" observedRunningTime="2025-09-30 14:11:34.713806101 +0000 UTC m=+923.342892554" watchObservedRunningTime="2025-09-30 14:11:34.715662329 +0000 UTC m=+923.344748762" Sep 30 14:11:34 crc kubenswrapper[4840]: I0930 14:11:34.756594 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=4.04333393 podStartE2EDuration="16.756534522s" podCreationTimestamp="2025-09-30 14:11:18 +0000 UTC" firstStartedPulling="2025-09-30 14:11:21.230174406 +0000 UTC m=+909.859260839" lastFinishedPulling="2025-09-30 14:11:33.943375008 +0000 UTC m=+922.572461431" observedRunningTime="2025-09-30 14:11:34.744074508 +0000 UTC m=+923.373160931" watchObservedRunningTime="2025-09-30 14:11:34.756534522 +0000 UTC m=+923.385620985" Sep 30 14:11:34 crc kubenswrapper[4840]: I0930 14:11:34.771858 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=9.72202408 podStartE2EDuration="12.77183689s" podCreationTimestamp="2025-09-30 14:11:22 +0000 UTC" firstStartedPulling="2025-09-30 14:11:30.914924364 +0000 UTC m=+919.544010787" lastFinishedPulling="2025-09-30 14:11:33.964737174 +0000 UTC m=+922.593823597" observedRunningTime="2025-09-30 14:11:34.768695208 +0000 UTC m=+923.397781641" watchObservedRunningTime="2025-09-30 14:11:34.77183689 +0000 UTC m=+923.400923323" Sep 30 14:11:34 crc kubenswrapper[4840]: I0930 14:11:34.794366 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-fc5jg" podStartSLOduration=10.319965683 podStartE2EDuration="15.794352525s" podCreationTimestamp="2025-09-30 14:11:19 +0000 UTC" firstStartedPulling="2025-09-30 14:11:24.931295701 +0000 UTC m=+913.560382124" lastFinishedPulling="2025-09-30 14:11:30.405682543 +0000 UTC m=+919.034768966" observedRunningTime="2025-09-30 14:11:34.792358373 +0000 UTC m=+923.421444796" watchObservedRunningTime="2025-09-30 14:11:34.794352525 +0000 UTC m=+923.423438948" Sep 30 14:11:34 crc kubenswrapper[4840]: I0930 14:11:34.822349 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7fd796d7df-wplqh" podStartSLOduration=12.822323992 podStartE2EDuration="12.822323992s" podCreationTimestamp="2025-09-30 14:11:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:11:34.812705322 +0000 UTC m=+923.441791745" watchObservedRunningTime="2025-09-30 14:11:34.822323992 +0000 UTC m=+923.451410405" Sep 30 14:11:34 crc kubenswrapper[4840]: I0930 14:11:34.993091 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-wplqh"] Sep 30 14:11:35 crc kubenswrapper[4840]: I0930 14:11:35.017661 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-gt8xx"] Sep 30 14:11:35 crc kubenswrapper[4840]: E0930 14:11:35.017954 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67556acd-23a4-40ab-b0ba-32c267b52956" containerName="dnsmasq-dns" Sep 30 14:11:35 crc kubenswrapper[4840]: I0930 14:11:35.017969 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="67556acd-23a4-40ab-b0ba-32c267b52956" containerName="dnsmasq-dns" Sep 30 14:11:35 crc kubenswrapper[4840]: E0930 14:11:35.017988 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67556acd-23a4-40ab-b0ba-32c267b52956" containerName="init" Sep 30 14:11:35 crc kubenswrapper[4840]: I0930 14:11:35.017995 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="67556acd-23a4-40ab-b0ba-32c267b52956" containerName="init" Sep 30 14:11:35 crc kubenswrapper[4840]: I0930 14:11:35.018143 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="67556acd-23a4-40ab-b0ba-32c267b52956" containerName="dnsmasq-dns" Sep 30 14:11:35 crc kubenswrapper[4840]: I0930 14:11:35.018928 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-gt8xx" Sep 30 14:11:35 crc kubenswrapper[4840]: I0930 14:11:35.021365 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Sep 30 14:11:35 crc kubenswrapper[4840]: I0930 14:11:35.032233 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-gt8xx"] Sep 30 14:11:35 crc kubenswrapper[4840]: I0930 14:11:35.096940 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Sep 30 14:11:35 crc kubenswrapper[4840]: I0930 14:11:35.096984 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Sep 30 14:11:35 crc kubenswrapper[4840]: I0930 14:11:35.113212 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f2759eed-4989-43b4-b025-4aab60dc864a-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-gt8xx\" (UID: \"f2759eed-4989-43b4-b025-4aab60dc864a\") " pod="openstack/dnsmasq-dns-86db49b7ff-gt8xx" Sep 30 14:11:35 crc kubenswrapper[4840]: I0930 14:11:35.113298 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2759eed-4989-43b4-b025-4aab60dc864a-config\") pod \"dnsmasq-dns-86db49b7ff-gt8xx\" (UID: \"f2759eed-4989-43b4-b025-4aab60dc864a\") " pod="openstack/dnsmasq-dns-86db49b7ff-gt8xx" Sep 30 14:11:35 crc kubenswrapper[4840]: I0930 14:11:35.113331 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtz4t\" (UniqueName: \"kubernetes.io/projected/f2759eed-4989-43b4-b025-4aab60dc864a-kube-api-access-mtz4t\") pod \"dnsmasq-dns-86db49b7ff-gt8xx\" (UID: \"f2759eed-4989-43b4-b025-4aab60dc864a\") " pod="openstack/dnsmasq-dns-86db49b7ff-gt8xx" Sep 30 14:11:35 crc kubenswrapper[4840]: I0930 14:11:35.113368 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f2759eed-4989-43b4-b025-4aab60dc864a-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-gt8xx\" (UID: \"f2759eed-4989-43b4-b025-4aab60dc864a\") " pod="openstack/dnsmasq-dns-86db49b7ff-gt8xx" Sep 30 14:11:35 crc kubenswrapper[4840]: I0930 14:11:35.113533 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f2759eed-4989-43b4-b025-4aab60dc864a-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-gt8xx\" (UID: \"f2759eed-4989-43b4-b025-4aab60dc864a\") " pod="openstack/dnsmasq-dns-86db49b7ff-gt8xx" Sep 30 14:11:35 crc kubenswrapper[4840]: I0930 14:11:35.132341 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Sep 30 14:11:35 crc kubenswrapper[4840]: I0930 14:11:35.215959 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f2759eed-4989-43b4-b025-4aab60dc864a-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-gt8xx\" (UID: \"f2759eed-4989-43b4-b025-4aab60dc864a\") " pod="openstack/dnsmasq-dns-86db49b7ff-gt8xx" Sep 30 14:11:35 crc kubenswrapper[4840]: I0930 14:11:35.216133 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2759eed-4989-43b4-b025-4aab60dc864a-config\") pod \"dnsmasq-dns-86db49b7ff-gt8xx\" (UID: \"f2759eed-4989-43b4-b025-4aab60dc864a\") " pod="openstack/dnsmasq-dns-86db49b7ff-gt8xx" Sep 30 14:11:35 crc kubenswrapper[4840]: I0930 14:11:35.216162 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtz4t\" (UniqueName: \"kubernetes.io/projected/f2759eed-4989-43b4-b025-4aab60dc864a-kube-api-access-mtz4t\") pod \"dnsmasq-dns-86db49b7ff-gt8xx\" (UID: \"f2759eed-4989-43b4-b025-4aab60dc864a\") " pod="openstack/dnsmasq-dns-86db49b7ff-gt8xx" Sep 30 14:11:35 crc kubenswrapper[4840]: I0930 14:11:35.216236 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f2759eed-4989-43b4-b025-4aab60dc864a-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-gt8xx\" (UID: \"f2759eed-4989-43b4-b025-4aab60dc864a\") " pod="openstack/dnsmasq-dns-86db49b7ff-gt8xx" Sep 30 14:11:35 crc kubenswrapper[4840]: I0930 14:11:35.216280 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f2759eed-4989-43b4-b025-4aab60dc864a-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-gt8xx\" (UID: \"f2759eed-4989-43b4-b025-4aab60dc864a\") " pod="openstack/dnsmasq-dns-86db49b7ff-gt8xx" Sep 30 14:11:35 crc kubenswrapper[4840]: I0930 14:11:35.216825 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f2759eed-4989-43b4-b025-4aab60dc864a-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-gt8xx\" (UID: \"f2759eed-4989-43b4-b025-4aab60dc864a\") " pod="openstack/dnsmasq-dns-86db49b7ff-gt8xx" Sep 30 14:11:35 crc kubenswrapper[4840]: I0930 14:11:35.217605 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f2759eed-4989-43b4-b025-4aab60dc864a-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-gt8xx\" (UID: \"f2759eed-4989-43b4-b025-4aab60dc864a\") " pod="openstack/dnsmasq-dns-86db49b7ff-gt8xx" Sep 30 14:11:35 crc kubenswrapper[4840]: I0930 14:11:35.217659 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f2759eed-4989-43b4-b025-4aab60dc864a-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-gt8xx\" (UID: \"f2759eed-4989-43b4-b025-4aab60dc864a\") " pod="openstack/dnsmasq-dns-86db49b7ff-gt8xx" Sep 30 14:11:35 crc kubenswrapper[4840]: I0930 14:11:35.217871 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2759eed-4989-43b4-b025-4aab60dc864a-config\") pod \"dnsmasq-dns-86db49b7ff-gt8xx\" (UID: \"f2759eed-4989-43b4-b025-4aab60dc864a\") " pod="openstack/dnsmasq-dns-86db49b7ff-gt8xx" Sep 30 14:11:35 crc kubenswrapper[4840]: I0930 14:11:35.230618 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-fc5jg" Sep 30 14:11:35 crc kubenswrapper[4840]: I0930 14:11:35.230655 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-fc5jg" Sep 30 14:11:35 crc kubenswrapper[4840]: I0930 14:11:35.239882 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtz4t\" (UniqueName: \"kubernetes.io/projected/f2759eed-4989-43b4-b025-4aab60dc864a-kube-api-access-mtz4t\") pod \"dnsmasq-dns-86db49b7ff-gt8xx\" (UID: \"f2759eed-4989-43b4-b025-4aab60dc864a\") " pod="openstack/dnsmasq-dns-86db49b7ff-gt8xx" Sep 30 14:11:35 crc kubenswrapper[4840]: I0930 14:11:35.342003 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-gt8xx" Sep 30 14:11:35 crc kubenswrapper[4840]: I0930 14:11:35.373264 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Sep 30 14:11:35 crc kubenswrapper[4840]: I0930 14:11:35.431214 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Sep 30 14:11:35 crc kubenswrapper[4840]: I0930 14:11:35.715090 4840 generic.go:334] "Generic (PLEG): container finished" podID="9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f" containerID="05ac69d0f33558ccad10e1fe8b72e19be47e55f4559138a238367ed19f65466d" exitCode=0 Sep 30 14:11:35 crc kubenswrapper[4840]: I0930 14:11:35.715229 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f","Type":"ContainerDied","Data":"05ac69d0f33558ccad10e1fe8b72e19be47e55f4559138a238367ed19f65466d"} Sep 30 14:11:35 crc kubenswrapper[4840]: I0930 14:11:35.717955 4840 generic.go:334] "Generic (PLEG): container finished" podID="41b0d67b-4ef4-4ebb-b000-b37bbdac6b34" containerID="b13172bf666fdd4d12af9ec2bb0f60e000d1d9ae612fc8995e6b5757136cfee3" exitCode=0 Sep 30 14:11:35 crc kubenswrapper[4840]: I0930 14:11:35.718850 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"41b0d67b-4ef4-4ebb-b000-b37bbdac6b34","Type":"ContainerDied","Data":"b13172bf666fdd4d12af9ec2bb0f60e000d1d9ae612fc8995e6b5757136cfee3"} Sep 30 14:11:35 crc kubenswrapper[4840]: I0930 14:11:35.719300 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Sep 30 14:11:35 crc kubenswrapper[4840]: I0930 14:11:35.783180 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-gt8xx"] Sep 30 14:11:35 crc kubenswrapper[4840]: W0930 14:11:35.785789 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf2759eed_4989_43b4_b025_4aab60dc864a.slice/crio-05bf60d61c0aca11f948fba1e90455fd5be9e1c7f953f6b9f16bbbf0551dab8d WatchSource:0}: Error finding container 05bf60d61c0aca11f948fba1e90455fd5be9e1c7f953f6b9f16bbbf0551dab8d: Status 404 returned error can't find the container with id 05bf60d61c0aca11f948fba1e90455fd5be9e1c7f953f6b9f16bbbf0551dab8d Sep 30 14:11:36 crc kubenswrapper[4840]: I0930 14:11:36.216971 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Sep 30 14:11:36 crc kubenswrapper[4840]: I0930 14:11:36.728217 4840 generic.go:334] "Generic (PLEG): container finished" podID="f2759eed-4989-43b4-b025-4aab60dc864a" containerID="4f71635a7189cd95395acdf5d262684f681fe8e3cb9c262b0c1d2ac9633e1d4a" exitCode=0 Sep 30 14:11:36 crc kubenswrapper[4840]: I0930 14:11:36.728464 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-gt8xx" event={"ID":"f2759eed-4989-43b4-b025-4aab60dc864a","Type":"ContainerDied","Data":"4f71635a7189cd95395acdf5d262684f681fe8e3cb9c262b0c1d2ac9633e1d4a"} Sep 30 14:11:36 crc kubenswrapper[4840]: I0930 14:11:36.728617 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-gt8xx" event={"ID":"f2759eed-4989-43b4-b025-4aab60dc864a","Type":"ContainerStarted","Data":"05bf60d61c0aca11f948fba1e90455fd5be9e1c7f953f6b9f16bbbf0551dab8d"} Sep 30 14:11:36 crc kubenswrapper[4840]: I0930 14:11:36.732911 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f","Type":"ContainerStarted","Data":"a7f6fd906e31c36b3d0554e6b9b88169f9f0b15d2ddd0d951d7dac6fce18e586"} Sep 30 14:11:36 crc kubenswrapper[4840]: I0930 14:11:36.735614 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"41b0d67b-4ef4-4ebb-b000-b37bbdac6b34","Type":"ContainerStarted","Data":"bfd54ef92d04ad68af9e1946c3612b631a2c6fce2eeae423dcc502491cf46c39"} Sep 30 14:11:36 crc kubenswrapper[4840]: I0930 14:11:36.736314 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7fd796d7df-wplqh" podUID="5cbf513e-7178-4a46-bc58-c3bfcfa728ff" containerName="dnsmasq-dns" containerID="cri-o://86e99e9ad660a3084b5ba24a7d04f4907f4359ad9abf610b2b14ab823d5d7a02" gracePeriod=10 Sep 30 14:11:36 crc kubenswrapper[4840]: I0930 14:11:36.784279 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=15.601721085 podStartE2EDuration="24.784258495s" podCreationTimestamp="2025-09-30 14:11:12 +0000 UTC" firstStartedPulling="2025-09-30 14:11:20.908286006 +0000 UTC m=+909.537372429" lastFinishedPulling="2025-09-30 14:11:30.090823416 +0000 UTC m=+918.719909839" observedRunningTime="2025-09-30 14:11:36.780138508 +0000 UTC m=+925.409224931" watchObservedRunningTime="2025-09-30 14:11:36.784258495 +0000 UTC m=+925.413344918" Sep 30 14:11:36 crc kubenswrapper[4840]: I0930 14:11:36.786544 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Sep 30 14:11:36 crc kubenswrapper[4840]: I0930 14:11:36.798746 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=15.207632499 podStartE2EDuration="24.798731062s" podCreationTimestamp="2025-09-30 14:11:12 +0000 UTC" firstStartedPulling="2025-09-30 14:11:20.920832013 +0000 UTC m=+909.549918436" lastFinishedPulling="2025-09-30 14:11:30.511930566 +0000 UTC m=+919.141016999" observedRunningTime="2025-09-30 14:11:36.797720225 +0000 UTC m=+925.426806648" watchObservedRunningTime="2025-09-30 14:11:36.798731062 +0000 UTC m=+925.427817485" Sep 30 14:11:37 crc kubenswrapper[4840]: I0930 14:11:37.159654 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-wplqh" Sep 30 14:11:37 crc kubenswrapper[4840]: I0930 14:11:37.250466 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5cbf513e-7178-4a46-bc58-c3bfcfa728ff-dns-svc\") pod \"5cbf513e-7178-4a46-bc58-c3bfcfa728ff\" (UID: \"5cbf513e-7178-4a46-bc58-c3bfcfa728ff\") " Sep 30 14:11:37 crc kubenswrapper[4840]: I0930 14:11:37.250586 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ksc66\" (UniqueName: \"kubernetes.io/projected/5cbf513e-7178-4a46-bc58-c3bfcfa728ff-kube-api-access-ksc66\") pod \"5cbf513e-7178-4a46-bc58-c3bfcfa728ff\" (UID: \"5cbf513e-7178-4a46-bc58-c3bfcfa728ff\") " Sep 30 14:11:37 crc kubenswrapper[4840]: I0930 14:11:37.250625 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cbf513e-7178-4a46-bc58-c3bfcfa728ff-config\") pod \"5cbf513e-7178-4a46-bc58-c3bfcfa728ff\" (UID: \"5cbf513e-7178-4a46-bc58-c3bfcfa728ff\") " Sep 30 14:11:37 crc kubenswrapper[4840]: I0930 14:11:37.250667 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5cbf513e-7178-4a46-bc58-c3bfcfa728ff-ovsdbserver-nb\") pod \"5cbf513e-7178-4a46-bc58-c3bfcfa728ff\" (UID: \"5cbf513e-7178-4a46-bc58-c3bfcfa728ff\") " Sep 30 14:11:37 crc kubenswrapper[4840]: I0930 14:11:37.271644 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cbf513e-7178-4a46-bc58-c3bfcfa728ff-kube-api-access-ksc66" (OuterVolumeSpecName: "kube-api-access-ksc66") pod "5cbf513e-7178-4a46-bc58-c3bfcfa728ff" (UID: "5cbf513e-7178-4a46-bc58-c3bfcfa728ff"). InnerVolumeSpecName "kube-api-access-ksc66". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:11:37 crc kubenswrapper[4840]: I0930 14:11:37.294311 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cbf513e-7178-4a46-bc58-c3bfcfa728ff-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5cbf513e-7178-4a46-bc58-c3bfcfa728ff" (UID: "5cbf513e-7178-4a46-bc58-c3bfcfa728ff"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:11:37 crc kubenswrapper[4840]: I0930 14:11:37.297100 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cbf513e-7178-4a46-bc58-c3bfcfa728ff-config" (OuterVolumeSpecName: "config") pod "5cbf513e-7178-4a46-bc58-c3bfcfa728ff" (UID: "5cbf513e-7178-4a46-bc58-c3bfcfa728ff"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:11:37 crc kubenswrapper[4840]: I0930 14:11:37.299663 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cbf513e-7178-4a46-bc58-c3bfcfa728ff-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5cbf513e-7178-4a46-bc58-c3bfcfa728ff" (UID: "5cbf513e-7178-4a46-bc58-c3bfcfa728ff"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:11:37 crc kubenswrapper[4840]: I0930 14:11:37.353931 4840 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5cbf513e-7178-4a46-bc58-c3bfcfa728ff-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 14:11:37 crc kubenswrapper[4840]: I0930 14:11:37.353990 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ksc66\" (UniqueName: \"kubernetes.io/projected/5cbf513e-7178-4a46-bc58-c3bfcfa728ff-kube-api-access-ksc66\") on node \"crc\" DevicePath \"\"" Sep 30 14:11:37 crc kubenswrapper[4840]: I0930 14:11:37.354008 4840 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cbf513e-7178-4a46-bc58-c3bfcfa728ff-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:11:37 crc kubenswrapper[4840]: I0930 14:11:37.354022 4840 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5cbf513e-7178-4a46-bc58-c3bfcfa728ff-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 14:11:37 crc kubenswrapper[4840]: I0930 14:11:37.745064 4840 generic.go:334] "Generic (PLEG): container finished" podID="5cbf513e-7178-4a46-bc58-c3bfcfa728ff" containerID="86e99e9ad660a3084b5ba24a7d04f4907f4359ad9abf610b2b14ab823d5d7a02" exitCode=0 Sep 30 14:11:37 crc kubenswrapper[4840]: I0930 14:11:37.745145 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-wplqh" event={"ID":"5cbf513e-7178-4a46-bc58-c3bfcfa728ff","Type":"ContainerDied","Data":"86e99e9ad660a3084b5ba24a7d04f4907f4359ad9abf610b2b14ab823d5d7a02"} Sep 30 14:11:37 crc kubenswrapper[4840]: I0930 14:11:37.745200 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-wplqh" event={"ID":"5cbf513e-7178-4a46-bc58-c3bfcfa728ff","Type":"ContainerDied","Data":"ebd48ad79841e44296b4299bccc3357e0379ab3822f7c6da68121734a42c4abd"} Sep 30 14:11:37 crc kubenswrapper[4840]: I0930 14:11:37.745218 4840 scope.go:117] "RemoveContainer" containerID="86e99e9ad660a3084b5ba24a7d04f4907f4359ad9abf610b2b14ab823d5d7a02" Sep 30 14:11:37 crc kubenswrapper[4840]: I0930 14:11:37.746397 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-wplqh" Sep 30 14:11:37 crc kubenswrapper[4840]: I0930 14:11:37.747298 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-gt8xx" event={"ID":"f2759eed-4989-43b4-b025-4aab60dc864a","Type":"ContainerStarted","Data":"81d2e9ea2e06c3dd73796140bb1862f4d42eeb2f12f035a1619c6007f2483bb7"} Sep 30 14:11:37 crc kubenswrapper[4840]: I0930 14:11:37.762398 4840 scope.go:117] "RemoveContainer" containerID="4f111f857791503b05a97de781e9f61a7de7f7f6f0420efd87a6878c46513080" Sep 30 14:11:37 crc kubenswrapper[4840]: I0930 14:11:37.769035 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-gt8xx" podStartSLOduration=3.769019521 podStartE2EDuration="3.769019521s" podCreationTimestamp="2025-09-30 14:11:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:11:37.76860623 +0000 UTC m=+926.397692673" watchObservedRunningTime="2025-09-30 14:11:37.769019521 +0000 UTC m=+926.398105944" Sep 30 14:11:37 crc kubenswrapper[4840]: I0930 14:11:37.800210 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-wplqh"] Sep 30 14:11:37 crc kubenswrapper[4840]: I0930 14:11:37.800954 4840 scope.go:117] "RemoveContainer" containerID="86e99e9ad660a3084b5ba24a7d04f4907f4359ad9abf610b2b14ab823d5d7a02" Sep 30 14:11:37 crc kubenswrapper[4840]: I0930 14:11:37.801661 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Sep 30 14:11:37 crc kubenswrapper[4840]: E0930 14:11:37.801895 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86e99e9ad660a3084b5ba24a7d04f4907f4359ad9abf610b2b14ab823d5d7a02\": container with ID starting with 86e99e9ad660a3084b5ba24a7d04f4907f4359ad9abf610b2b14ab823d5d7a02 not found: ID does not exist" containerID="86e99e9ad660a3084b5ba24a7d04f4907f4359ad9abf610b2b14ab823d5d7a02" Sep 30 14:11:37 crc kubenswrapper[4840]: I0930 14:11:37.801999 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86e99e9ad660a3084b5ba24a7d04f4907f4359ad9abf610b2b14ab823d5d7a02"} err="failed to get container status \"86e99e9ad660a3084b5ba24a7d04f4907f4359ad9abf610b2b14ab823d5d7a02\": rpc error: code = NotFound desc = could not find container \"86e99e9ad660a3084b5ba24a7d04f4907f4359ad9abf610b2b14ab823d5d7a02\": container with ID starting with 86e99e9ad660a3084b5ba24a7d04f4907f4359ad9abf610b2b14ab823d5d7a02 not found: ID does not exist" Sep 30 14:11:37 crc kubenswrapper[4840]: I0930 14:11:37.802098 4840 scope.go:117] "RemoveContainer" containerID="4f111f857791503b05a97de781e9f61a7de7f7f6f0420efd87a6878c46513080" Sep 30 14:11:37 crc kubenswrapper[4840]: E0930 14:11:37.802973 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f111f857791503b05a97de781e9f61a7de7f7f6f0420efd87a6878c46513080\": container with ID starting with 4f111f857791503b05a97de781e9f61a7de7f7f6f0420efd87a6878c46513080 not found: ID does not exist" containerID="4f111f857791503b05a97de781e9f61a7de7f7f6f0420efd87a6878c46513080" Sep 30 14:11:37 crc kubenswrapper[4840]: I0930 14:11:37.803070 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f111f857791503b05a97de781e9f61a7de7f7f6f0420efd87a6878c46513080"} err="failed to get container status \"4f111f857791503b05a97de781e9f61a7de7f7f6f0420efd87a6878c46513080\": rpc error: code = NotFound desc = could not find container \"4f111f857791503b05a97de781e9f61a7de7f7f6f0420efd87a6878c46513080\": container with ID starting with 4f111f857791503b05a97de781e9f61a7de7f7f6f0420efd87a6878c46513080 not found: ID does not exist" Sep 30 14:11:37 crc kubenswrapper[4840]: I0930 14:11:37.809337 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-wplqh"] Sep 30 14:11:37 crc kubenswrapper[4840]: I0930 14:11:37.934823 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Sep 30 14:11:37 crc kubenswrapper[4840]: E0930 14:11:37.935312 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cbf513e-7178-4a46-bc58-c3bfcfa728ff" containerName="dnsmasq-dns" Sep 30 14:11:37 crc kubenswrapper[4840]: I0930 14:11:37.935928 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cbf513e-7178-4a46-bc58-c3bfcfa728ff" containerName="dnsmasq-dns" Sep 30 14:11:37 crc kubenswrapper[4840]: E0930 14:11:37.936003 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cbf513e-7178-4a46-bc58-c3bfcfa728ff" containerName="init" Sep 30 14:11:37 crc kubenswrapper[4840]: I0930 14:11:37.936064 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cbf513e-7178-4a46-bc58-c3bfcfa728ff" containerName="init" Sep 30 14:11:37 crc kubenswrapper[4840]: I0930 14:11:37.936248 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cbf513e-7178-4a46-bc58-c3bfcfa728ff" containerName="dnsmasq-dns" Sep 30 14:11:37 crc kubenswrapper[4840]: I0930 14:11:37.937122 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Sep 30 14:11:37 crc kubenswrapper[4840]: I0930 14:11:37.939650 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Sep 30 14:11:37 crc kubenswrapper[4840]: I0930 14:11:37.941550 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-mnfmz" Sep 30 14:11:37 crc kubenswrapper[4840]: I0930 14:11:37.941558 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Sep 30 14:11:37 crc kubenswrapper[4840]: I0930 14:11:37.941550 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Sep 30 14:11:37 crc kubenswrapper[4840]: I0930 14:11:37.945781 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Sep 30 14:11:38 crc kubenswrapper[4840]: I0930 14:11:38.066220 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2338a6b1-3726-4994-9325-ee16f93e28c0-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"2338a6b1-3726-4994-9325-ee16f93e28c0\") " pod="openstack/ovn-northd-0" Sep 30 14:11:38 crc kubenswrapper[4840]: I0930 14:11:38.066527 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2338a6b1-3726-4994-9325-ee16f93e28c0-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"2338a6b1-3726-4994-9325-ee16f93e28c0\") " pod="openstack/ovn-northd-0" Sep 30 14:11:38 crc kubenswrapper[4840]: I0930 14:11:38.066561 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/2338a6b1-3726-4994-9325-ee16f93e28c0-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"2338a6b1-3726-4994-9325-ee16f93e28c0\") " pod="openstack/ovn-northd-0" Sep 30 14:11:38 crc kubenswrapper[4840]: I0930 14:11:38.066634 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2338a6b1-3726-4994-9325-ee16f93e28c0-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"2338a6b1-3726-4994-9325-ee16f93e28c0\") " pod="openstack/ovn-northd-0" Sep 30 14:11:38 crc kubenswrapper[4840]: I0930 14:11:38.066715 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2338a6b1-3726-4994-9325-ee16f93e28c0-config\") pod \"ovn-northd-0\" (UID: \"2338a6b1-3726-4994-9325-ee16f93e28c0\") " pod="openstack/ovn-northd-0" Sep 30 14:11:38 crc kubenswrapper[4840]: I0930 14:11:38.066805 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2338a6b1-3726-4994-9325-ee16f93e28c0-scripts\") pod \"ovn-northd-0\" (UID: \"2338a6b1-3726-4994-9325-ee16f93e28c0\") " pod="openstack/ovn-northd-0" Sep 30 14:11:38 crc kubenswrapper[4840]: I0930 14:11:38.066830 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2r2t\" (UniqueName: \"kubernetes.io/projected/2338a6b1-3726-4994-9325-ee16f93e28c0-kube-api-access-c2r2t\") pod \"ovn-northd-0\" (UID: \"2338a6b1-3726-4994-9325-ee16f93e28c0\") " pod="openstack/ovn-northd-0" Sep 30 14:11:38 crc kubenswrapper[4840]: I0930 14:11:38.132890 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cbf513e-7178-4a46-bc58-c3bfcfa728ff" path="/var/lib/kubelet/pods/5cbf513e-7178-4a46-bc58-c3bfcfa728ff/volumes" Sep 30 14:11:38 crc kubenswrapper[4840]: I0930 14:11:38.167675 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2338a6b1-3726-4994-9325-ee16f93e28c0-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"2338a6b1-3726-4994-9325-ee16f93e28c0\") " pod="openstack/ovn-northd-0" Sep 30 14:11:38 crc kubenswrapper[4840]: I0930 14:11:38.167731 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2338a6b1-3726-4994-9325-ee16f93e28c0-config\") pod \"ovn-northd-0\" (UID: \"2338a6b1-3726-4994-9325-ee16f93e28c0\") " pod="openstack/ovn-northd-0" Sep 30 14:11:38 crc kubenswrapper[4840]: I0930 14:11:38.167784 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2338a6b1-3726-4994-9325-ee16f93e28c0-scripts\") pod \"ovn-northd-0\" (UID: \"2338a6b1-3726-4994-9325-ee16f93e28c0\") " pod="openstack/ovn-northd-0" Sep 30 14:11:38 crc kubenswrapper[4840]: I0930 14:11:38.167804 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2r2t\" (UniqueName: \"kubernetes.io/projected/2338a6b1-3726-4994-9325-ee16f93e28c0-kube-api-access-c2r2t\") pod \"ovn-northd-0\" (UID: \"2338a6b1-3726-4994-9325-ee16f93e28c0\") " pod="openstack/ovn-northd-0" Sep 30 14:11:38 crc kubenswrapper[4840]: I0930 14:11:38.167849 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2338a6b1-3726-4994-9325-ee16f93e28c0-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"2338a6b1-3726-4994-9325-ee16f93e28c0\") " pod="openstack/ovn-northd-0" Sep 30 14:11:38 crc kubenswrapper[4840]: I0930 14:11:38.167869 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2338a6b1-3726-4994-9325-ee16f93e28c0-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"2338a6b1-3726-4994-9325-ee16f93e28c0\") " pod="openstack/ovn-northd-0" Sep 30 14:11:38 crc kubenswrapper[4840]: I0930 14:11:38.167887 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/2338a6b1-3726-4994-9325-ee16f93e28c0-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"2338a6b1-3726-4994-9325-ee16f93e28c0\") " pod="openstack/ovn-northd-0" Sep 30 14:11:38 crc kubenswrapper[4840]: I0930 14:11:38.168982 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2338a6b1-3726-4994-9325-ee16f93e28c0-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"2338a6b1-3726-4994-9325-ee16f93e28c0\") " pod="openstack/ovn-northd-0" Sep 30 14:11:38 crc kubenswrapper[4840]: I0930 14:11:38.169351 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2338a6b1-3726-4994-9325-ee16f93e28c0-config\") pod \"ovn-northd-0\" (UID: \"2338a6b1-3726-4994-9325-ee16f93e28c0\") " pod="openstack/ovn-northd-0" Sep 30 14:11:38 crc kubenswrapper[4840]: I0930 14:11:38.170684 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2338a6b1-3726-4994-9325-ee16f93e28c0-scripts\") pod \"ovn-northd-0\" (UID: \"2338a6b1-3726-4994-9325-ee16f93e28c0\") " pod="openstack/ovn-northd-0" Sep 30 14:11:38 crc kubenswrapper[4840]: I0930 14:11:38.171917 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2338a6b1-3726-4994-9325-ee16f93e28c0-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"2338a6b1-3726-4994-9325-ee16f93e28c0\") " pod="openstack/ovn-northd-0" Sep 30 14:11:38 crc kubenswrapper[4840]: I0930 14:11:38.172243 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/2338a6b1-3726-4994-9325-ee16f93e28c0-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"2338a6b1-3726-4994-9325-ee16f93e28c0\") " pod="openstack/ovn-northd-0" Sep 30 14:11:38 crc kubenswrapper[4840]: I0930 14:11:38.173775 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2338a6b1-3726-4994-9325-ee16f93e28c0-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"2338a6b1-3726-4994-9325-ee16f93e28c0\") " pod="openstack/ovn-northd-0" Sep 30 14:11:38 crc kubenswrapper[4840]: I0930 14:11:38.188033 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2r2t\" (UniqueName: \"kubernetes.io/projected/2338a6b1-3726-4994-9325-ee16f93e28c0-kube-api-access-c2r2t\") pod \"ovn-northd-0\" (UID: \"2338a6b1-3726-4994-9325-ee16f93e28c0\") " pod="openstack/ovn-northd-0" Sep 30 14:11:38 crc kubenswrapper[4840]: I0930 14:11:38.256634 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Sep 30 14:11:38 crc kubenswrapper[4840]: I0930 14:11:38.663466 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Sep 30 14:11:38 crc kubenswrapper[4840]: W0930 14:11:38.669483 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2338a6b1_3726_4994_9325_ee16f93e28c0.slice/crio-cc661afba76550e94601f09d46d9afccd9de2dcc77001a4e4f1f3dce8672e0bf WatchSource:0}: Error finding container cc661afba76550e94601f09d46d9afccd9de2dcc77001a4e4f1f3dce8672e0bf: Status 404 returned error can't find the container with id cc661afba76550e94601f09d46d9afccd9de2dcc77001a4e4f1f3dce8672e0bf Sep 30 14:11:38 crc kubenswrapper[4840]: I0930 14:11:38.754576 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"2338a6b1-3726-4994-9325-ee16f93e28c0","Type":"ContainerStarted","Data":"cc661afba76550e94601f09d46d9afccd9de2dcc77001a4e4f1f3dce8672e0bf"} Sep 30 14:11:38 crc kubenswrapper[4840]: I0930 14:11:38.756770 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-gt8xx" Sep 30 14:11:39 crc kubenswrapper[4840]: I0930 14:11:39.167812 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Sep 30 14:11:40 crc kubenswrapper[4840]: I0930 14:11:40.774223 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"2338a6b1-3726-4994-9325-ee16f93e28c0","Type":"ContainerStarted","Data":"d102b6c8ebddc7e5ff92e4a4eb88f3e62f9396e30495bfd83dbaa790b512e760"} Sep 30 14:11:40 crc kubenswrapper[4840]: I0930 14:11:40.774271 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"2338a6b1-3726-4994-9325-ee16f93e28c0","Type":"ContainerStarted","Data":"eb9feb1cb1f5622f98e5698110170005e49eb2d8bee1372253eb7161de9bdeb0"} Sep 30 14:11:40 crc kubenswrapper[4840]: I0930 14:11:40.774377 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Sep 30 14:11:40 crc kubenswrapper[4840]: I0930 14:11:40.794829 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.707967718 podStartE2EDuration="3.794806096s" podCreationTimestamp="2025-09-30 14:11:37 +0000 UTC" firstStartedPulling="2025-09-30 14:11:38.673448868 +0000 UTC m=+927.302535291" lastFinishedPulling="2025-09-30 14:11:39.760287246 +0000 UTC m=+928.389373669" observedRunningTime="2025-09-30 14:11:40.789229791 +0000 UTC m=+929.418316234" watchObservedRunningTime="2025-09-30 14:11:40.794806096 +0000 UTC m=+929.423892519" Sep 30 14:11:43 crc kubenswrapper[4840]: I0930 14:11:43.782537 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Sep 30 14:11:43 crc kubenswrapper[4840]: I0930 14:11:43.782613 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Sep 30 14:11:43 crc kubenswrapper[4840]: I0930 14:11:43.964486 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Sep 30 14:11:43 crc kubenswrapper[4840]: I0930 14:11:43.964943 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Sep 30 14:11:44 crc kubenswrapper[4840]: I0930 14:11:44.034422 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Sep 30 14:11:44 crc kubenswrapper[4840]: I0930 14:11:44.869070 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Sep 30 14:11:45 crc kubenswrapper[4840]: I0930 14:11:45.344628 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-gt8xx" Sep 30 14:11:45 crc kubenswrapper[4840]: I0930 14:11:45.393911 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-pvrll"] Sep 30 14:11:45 crc kubenswrapper[4840]: I0930 14:11:45.394181 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5ccc8479f9-pvrll" podUID="9f2e45b7-fe3e-4e53-a735-05d843a14159" containerName="dnsmasq-dns" containerID="cri-o://b6817e74a280d4642bbb241ea913659cb63427fbe679a334fe8dfdf5f1a1188d" gracePeriod=10 Sep 30 14:11:45 crc kubenswrapper[4840]: I0930 14:11:45.813859 4840 generic.go:334] "Generic (PLEG): container finished" podID="9f2e45b7-fe3e-4e53-a735-05d843a14159" containerID="b6817e74a280d4642bbb241ea913659cb63427fbe679a334fe8dfdf5f1a1188d" exitCode=0 Sep 30 14:11:45 crc kubenswrapper[4840]: I0930 14:11:45.813926 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-pvrll" event={"ID":"9f2e45b7-fe3e-4e53-a735-05d843a14159","Type":"ContainerDied","Data":"b6817e74a280d4642bbb241ea913659cb63427fbe679a334fe8dfdf5f1a1188d"} Sep 30 14:11:45 crc kubenswrapper[4840]: I0930 14:11:45.813991 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-pvrll" event={"ID":"9f2e45b7-fe3e-4e53-a735-05d843a14159","Type":"ContainerDied","Data":"9fe96403f9dbc4f3c79eccdcece4ddf38c69fdec2c7499238253c979ed1b99bd"} Sep 30 14:11:45 crc kubenswrapper[4840]: I0930 14:11:45.814010 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9fe96403f9dbc4f3c79eccdcece4ddf38c69fdec2c7499238253c979ed1b99bd" Sep 30 14:11:45 crc kubenswrapper[4840]: I0930 14:11:45.831993 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-pvrll" Sep 30 14:11:45 crc kubenswrapper[4840]: I0930 14:11:45.843274 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Sep 30 14:11:45 crc kubenswrapper[4840]: I0930 14:11:45.889228 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f2e45b7-fe3e-4e53-a735-05d843a14159-dns-svc\") pod \"9f2e45b7-fe3e-4e53-a735-05d843a14159\" (UID: \"9f2e45b7-fe3e-4e53-a735-05d843a14159\") " Sep 30 14:11:45 crc kubenswrapper[4840]: I0930 14:11:45.889371 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qvr9\" (UniqueName: \"kubernetes.io/projected/9f2e45b7-fe3e-4e53-a735-05d843a14159-kube-api-access-9qvr9\") pod \"9f2e45b7-fe3e-4e53-a735-05d843a14159\" (UID: \"9f2e45b7-fe3e-4e53-a735-05d843a14159\") " Sep 30 14:11:45 crc kubenswrapper[4840]: I0930 14:11:45.889424 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f2e45b7-fe3e-4e53-a735-05d843a14159-config\") pod \"9f2e45b7-fe3e-4e53-a735-05d843a14159\" (UID: \"9f2e45b7-fe3e-4e53-a735-05d843a14159\") " Sep 30 14:11:45 crc kubenswrapper[4840]: I0930 14:11:45.900528 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Sep 30 14:11:45 crc kubenswrapper[4840]: I0930 14:11:45.904343 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f2e45b7-fe3e-4e53-a735-05d843a14159-kube-api-access-9qvr9" (OuterVolumeSpecName: "kube-api-access-9qvr9") pod "9f2e45b7-fe3e-4e53-a735-05d843a14159" (UID: "9f2e45b7-fe3e-4e53-a735-05d843a14159"). InnerVolumeSpecName "kube-api-access-9qvr9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:11:45 crc kubenswrapper[4840]: I0930 14:11:45.928955 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f2e45b7-fe3e-4e53-a735-05d843a14159-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9f2e45b7-fe3e-4e53-a735-05d843a14159" (UID: "9f2e45b7-fe3e-4e53-a735-05d843a14159"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:11:45 crc kubenswrapper[4840]: I0930 14:11:45.935965 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f2e45b7-fe3e-4e53-a735-05d843a14159-config" (OuterVolumeSpecName: "config") pod "9f2e45b7-fe3e-4e53-a735-05d843a14159" (UID: "9f2e45b7-fe3e-4e53-a735-05d843a14159"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:11:45 crc kubenswrapper[4840]: I0930 14:11:45.992410 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qvr9\" (UniqueName: \"kubernetes.io/projected/9f2e45b7-fe3e-4e53-a735-05d843a14159-kube-api-access-9qvr9\") on node \"crc\" DevicePath \"\"" Sep 30 14:11:45 crc kubenswrapper[4840]: I0930 14:11:45.992449 4840 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f2e45b7-fe3e-4e53-a735-05d843a14159-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:11:45 crc kubenswrapper[4840]: I0930 14:11:45.992458 4840 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f2e45b7-fe3e-4e53-a735-05d843a14159-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 14:11:46 crc kubenswrapper[4840]: I0930 14:11:46.206682 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-2mqrn"] Sep 30 14:11:46 crc kubenswrapper[4840]: E0930 14:11:46.206967 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f2e45b7-fe3e-4e53-a735-05d843a14159" containerName="dnsmasq-dns" Sep 30 14:11:46 crc kubenswrapper[4840]: I0930 14:11:46.206983 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f2e45b7-fe3e-4e53-a735-05d843a14159" containerName="dnsmasq-dns" Sep 30 14:11:46 crc kubenswrapper[4840]: E0930 14:11:46.206996 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f2e45b7-fe3e-4e53-a735-05d843a14159" containerName="init" Sep 30 14:11:46 crc kubenswrapper[4840]: I0930 14:11:46.207002 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f2e45b7-fe3e-4e53-a735-05d843a14159" containerName="init" Sep 30 14:11:46 crc kubenswrapper[4840]: I0930 14:11:46.207185 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f2e45b7-fe3e-4e53-a735-05d843a14159" containerName="dnsmasq-dns" Sep 30 14:11:46 crc kubenswrapper[4840]: I0930 14:11:46.208001 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-2mqrn" Sep 30 14:11:46 crc kubenswrapper[4840]: I0930 14:11:46.228638 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-2mqrn"] Sep 30 14:11:46 crc kubenswrapper[4840]: I0930 14:11:46.296764 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/377cf7b4-32f7-4aa9-85f2-1b62763d9562-dns-svc\") pod \"dnsmasq-dns-698758b865-2mqrn\" (UID: \"377cf7b4-32f7-4aa9-85f2-1b62763d9562\") " pod="openstack/dnsmasq-dns-698758b865-2mqrn" Sep 30 14:11:46 crc kubenswrapper[4840]: I0930 14:11:46.296837 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/377cf7b4-32f7-4aa9-85f2-1b62763d9562-config\") pod \"dnsmasq-dns-698758b865-2mqrn\" (UID: \"377cf7b4-32f7-4aa9-85f2-1b62763d9562\") " pod="openstack/dnsmasq-dns-698758b865-2mqrn" Sep 30 14:11:46 crc kubenswrapper[4840]: I0930 14:11:46.296892 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/377cf7b4-32f7-4aa9-85f2-1b62763d9562-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-2mqrn\" (UID: \"377cf7b4-32f7-4aa9-85f2-1b62763d9562\") " pod="openstack/dnsmasq-dns-698758b865-2mqrn" Sep 30 14:11:46 crc kubenswrapper[4840]: I0930 14:11:46.296919 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhmrq\" (UniqueName: \"kubernetes.io/projected/377cf7b4-32f7-4aa9-85f2-1b62763d9562-kube-api-access-mhmrq\") pod \"dnsmasq-dns-698758b865-2mqrn\" (UID: \"377cf7b4-32f7-4aa9-85f2-1b62763d9562\") " pod="openstack/dnsmasq-dns-698758b865-2mqrn" Sep 30 14:11:46 crc kubenswrapper[4840]: I0930 14:11:46.296985 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/377cf7b4-32f7-4aa9-85f2-1b62763d9562-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-2mqrn\" (UID: \"377cf7b4-32f7-4aa9-85f2-1b62763d9562\") " pod="openstack/dnsmasq-dns-698758b865-2mqrn" Sep 30 14:11:46 crc kubenswrapper[4840]: I0930 14:11:46.397817 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/377cf7b4-32f7-4aa9-85f2-1b62763d9562-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-2mqrn\" (UID: \"377cf7b4-32f7-4aa9-85f2-1b62763d9562\") " pod="openstack/dnsmasq-dns-698758b865-2mqrn" Sep 30 14:11:46 crc kubenswrapper[4840]: I0930 14:11:46.397867 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhmrq\" (UniqueName: \"kubernetes.io/projected/377cf7b4-32f7-4aa9-85f2-1b62763d9562-kube-api-access-mhmrq\") pod \"dnsmasq-dns-698758b865-2mqrn\" (UID: \"377cf7b4-32f7-4aa9-85f2-1b62763d9562\") " pod="openstack/dnsmasq-dns-698758b865-2mqrn" Sep 30 14:11:46 crc kubenswrapper[4840]: I0930 14:11:46.397940 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/377cf7b4-32f7-4aa9-85f2-1b62763d9562-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-2mqrn\" (UID: \"377cf7b4-32f7-4aa9-85f2-1b62763d9562\") " pod="openstack/dnsmasq-dns-698758b865-2mqrn" Sep 30 14:11:46 crc kubenswrapper[4840]: I0930 14:11:46.397969 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/377cf7b4-32f7-4aa9-85f2-1b62763d9562-dns-svc\") pod \"dnsmasq-dns-698758b865-2mqrn\" (UID: \"377cf7b4-32f7-4aa9-85f2-1b62763d9562\") " pod="openstack/dnsmasq-dns-698758b865-2mqrn" Sep 30 14:11:46 crc kubenswrapper[4840]: I0930 14:11:46.398038 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/377cf7b4-32f7-4aa9-85f2-1b62763d9562-config\") pod \"dnsmasq-dns-698758b865-2mqrn\" (UID: \"377cf7b4-32f7-4aa9-85f2-1b62763d9562\") " pod="openstack/dnsmasq-dns-698758b865-2mqrn" Sep 30 14:11:46 crc kubenswrapper[4840]: I0930 14:11:46.399161 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/377cf7b4-32f7-4aa9-85f2-1b62763d9562-dns-svc\") pod \"dnsmasq-dns-698758b865-2mqrn\" (UID: \"377cf7b4-32f7-4aa9-85f2-1b62763d9562\") " pod="openstack/dnsmasq-dns-698758b865-2mqrn" Sep 30 14:11:46 crc kubenswrapper[4840]: I0930 14:11:46.399187 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/377cf7b4-32f7-4aa9-85f2-1b62763d9562-config\") pod \"dnsmasq-dns-698758b865-2mqrn\" (UID: \"377cf7b4-32f7-4aa9-85f2-1b62763d9562\") " pod="openstack/dnsmasq-dns-698758b865-2mqrn" Sep 30 14:11:46 crc kubenswrapper[4840]: I0930 14:11:46.399265 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/377cf7b4-32f7-4aa9-85f2-1b62763d9562-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-2mqrn\" (UID: \"377cf7b4-32f7-4aa9-85f2-1b62763d9562\") " pod="openstack/dnsmasq-dns-698758b865-2mqrn" Sep 30 14:11:46 crc kubenswrapper[4840]: I0930 14:11:46.399287 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/377cf7b4-32f7-4aa9-85f2-1b62763d9562-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-2mqrn\" (UID: \"377cf7b4-32f7-4aa9-85f2-1b62763d9562\") " pod="openstack/dnsmasq-dns-698758b865-2mqrn" Sep 30 14:11:46 crc kubenswrapper[4840]: I0930 14:11:46.413752 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhmrq\" (UniqueName: \"kubernetes.io/projected/377cf7b4-32f7-4aa9-85f2-1b62763d9562-kube-api-access-mhmrq\") pod \"dnsmasq-dns-698758b865-2mqrn\" (UID: \"377cf7b4-32f7-4aa9-85f2-1b62763d9562\") " pod="openstack/dnsmasq-dns-698758b865-2mqrn" Sep 30 14:11:46 crc kubenswrapper[4840]: I0930 14:11:46.529788 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-2mqrn" Sep 30 14:11:46 crc kubenswrapper[4840]: I0930 14:11:46.843757 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-pvrll" Sep 30 14:11:46 crc kubenswrapper[4840]: I0930 14:11:46.875296 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-pvrll"] Sep 30 14:11:46 crc kubenswrapper[4840]: I0930 14:11:46.881176 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-pvrll"] Sep 30 14:11:46 crc kubenswrapper[4840]: I0930 14:11:46.942206 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-2mqrn"] Sep 30 14:11:46 crc kubenswrapper[4840]: W0930 14:11:46.947473 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod377cf7b4_32f7_4aa9_85f2_1b62763d9562.slice/crio-93adcd8b5e6c5f908293036f453c8596b3a9519a62b6fe0fa82a852d7d8d25b1 WatchSource:0}: Error finding container 93adcd8b5e6c5f908293036f453c8596b3a9519a62b6fe0fa82a852d7d8d25b1: Status 404 returned error can't find the container with id 93adcd8b5e6c5f908293036f453c8596b3a9519a62b6fe0fa82a852d7d8d25b1 Sep 30 14:11:47 crc kubenswrapper[4840]: I0930 14:11:47.339401 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Sep 30 14:11:47 crc kubenswrapper[4840]: I0930 14:11:47.344791 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Sep 30 14:11:47 crc kubenswrapper[4840]: I0930 14:11:47.346383 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Sep 30 14:11:47 crc kubenswrapper[4840]: I0930 14:11:47.346806 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-hvktb" Sep 30 14:11:47 crc kubenswrapper[4840]: I0930 14:11:47.347432 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Sep 30 14:11:47 crc kubenswrapper[4840]: I0930 14:11:47.347488 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Sep 30 14:11:47 crc kubenswrapper[4840]: I0930 14:11:47.371409 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Sep 30 14:11:47 crc kubenswrapper[4840]: I0930 14:11:47.426400 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/66a7a4e6-dd49-4589-bd88-2f5fdf0af70f-etc-swift\") pod \"swift-storage-0\" (UID: \"66a7a4e6-dd49-4589-bd88-2f5fdf0af70f\") " pod="openstack/swift-storage-0" Sep 30 14:11:47 crc kubenswrapper[4840]: I0930 14:11:47.426456 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8v6d\" (UniqueName: \"kubernetes.io/projected/66a7a4e6-dd49-4589-bd88-2f5fdf0af70f-kube-api-access-n8v6d\") pod \"swift-storage-0\" (UID: \"66a7a4e6-dd49-4589-bd88-2f5fdf0af70f\") " pod="openstack/swift-storage-0" Sep 30 14:11:47 crc kubenswrapper[4840]: I0930 14:11:47.426539 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/66a7a4e6-dd49-4589-bd88-2f5fdf0af70f-cache\") pod \"swift-storage-0\" (UID: \"66a7a4e6-dd49-4589-bd88-2f5fdf0af70f\") " pod="openstack/swift-storage-0" Sep 30 14:11:47 crc kubenswrapper[4840]: I0930 14:11:47.426579 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"66a7a4e6-dd49-4589-bd88-2f5fdf0af70f\") " pod="openstack/swift-storage-0" Sep 30 14:11:47 crc kubenswrapper[4840]: I0930 14:11:47.426618 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/66a7a4e6-dd49-4589-bd88-2f5fdf0af70f-lock\") pod \"swift-storage-0\" (UID: \"66a7a4e6-dd49-4589-bd88-2f5fdf0af70f\") " pod="openstack/swift-storage-0" Sep 30 14:11:47 crc kubenswrapper[4840]: I0930 14:11:47.528168 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/66a7a4e6-dd49-4589-bd88-2f5fdf0af70f-lock\") pod \"swift-storage-0\" (UID: \"66a7a4e6-dd49-4589-bd88-2f5fdf0af70f\") " pod="openstack/swift-storage-0" Sep 30 14:11:47 crc kubenswrapper[4840]: I0930 14:11:47.528262 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/66a7a4e6-dd49-4589-bd88-2f5fdf0af70f-etc-swift\") pod \"swift-storage-0\" (UID: \"66a7a4e6-dd49-4589-bd88-2f5fdf0af70f\") " pod="openstack/swift-storage-0" Sep 30 14:11:47 crc kubenswrapper[4840]: I0930 14:11:47.528290 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8v6d\" (UniqueName: \"kubernetes.io/projected/66a7a4e6-dd49-4589-bd88-2f5fdf0af70f-kube-api-access-n8v6d\") pod \"swift-storage-0\" (UID: \"66a7a4e6-dd49-4589-bd88-2f5fdf0af70f\") " pod="openstack/swift-storage-0" Sep 30 14:11:47 crc kubenswrapper[4840]: I0930 14:11:47.528348 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/66a7a4e6-dd49-4589-bd88-2f5fdf0af70f-cache\") pod \"swift-storage-0\" (UID: \"66a7a4e6-dd49-4589-bd88-2f5fdf0af70f\") " pod="openstack/swift-storage-0" Sep 30 14:11:47 crc kubenswrapper[4840]: I0930 14:11:47.528370 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"66a7a4e6-dd49-4589-bd88-2f5fdf0af70f\") " pod="openstack/swift-storage-0" Sep 30 14:11:47 crc kubenswrapper[4840]: E0930 14:11:47.528496 4840 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Sep 30 14:11:47 crc kubenswrapper[4840]: E0930 14:11:47.528537 4840 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Sep 30 14:11:47 crc kubenswrapper[4840]: I0930 14:11:47.528669 4840 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"66a7a4e6-dd49-4589-bd88-2f5fdf0af70f\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/swift-storage-0" Sep 30 14:11:47 crc kubenswrapper[4840]: E0930 14:11:47.528673 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/66a7a4e6-dd49-4589-bd88-2f5fdf0af70f-etc-swift podName:66a7a4e6-dd49-4589-bd88-2f5fdf0af70f nodeName:}" failed. No retries permitted until 2025-09-30 14:11:48.028653028 +0000 UTC m=+936.657739461 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/66a7a4e6-dd49-4589-bd88-2f5fdf0af70f-etc-swift") pod "swift-storage-0" (UID: "66a7a4e6-dd49-4589-bd88-2f5fdf0af70f") : configmap "swift-ring-files" not found Sep 30 14:11:47 crc kubenswrapper[4840]: I0930 14:11:47.528802 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/66a7a4e6-dd49-4589-bd88-2f5fdf0af70f-lock\") pod \"swift-storage-0\" (UID: \"66a7a4e6-dd49-4589-bd88-2f5fdf0af70f\") " pod="openstack/swift-storage-0" Sep 30 14:11:47 crc kubenswrapper[4840]: I0930 14:11:47.529446 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/66a7a4e6-dd49-4589-bd88-2f5fdf0af70f-cache\") pod \"swift-storage-0\" (UID: \"66a7a4e6-dd49-4589-bd88-2f5fdf0af70f\") " pod="openstack/swift-storage-0" Sep 30 14:11:47 crc kubenswrapper[4840]: I0930 14:11:47.545336 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8v6d\" (UniqueName: \"kubernetes.io/projected/66a7a4e6-dd49-4589-bd88-2f5fdf0af70f-kube-api-access-n8v6d\") pod \"swift-storage-0\" (UID: \"66a7a4e6-dd49-4589-bd88-2f5fdf0af70f\") " pod="openstack/swift-storage-0" Sep 30 14:11:47 crc kubenswrapper[4840]: I0930 14:11:47.548188 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"66a7a4e6-dd49-4589-bd88-2f5fdf0af70f\") " pod="openstack/swift-storage-0" Sep 30 14:11:47 crc kubenswrapper[4840]: I0930 14:11:47.846198 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-txwn9"] Sep 30 14:11:47 crc kubenswrapper[4840]: I0930 14:11:47.847318 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-txwn9" Sep 30 14:11:47 crc kubenswrapper[4840]: I0930 14:11:47.850109 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Sep 30 14:11:47 crc kubenswrapper[4840]: I0930 14:11:47.850194 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Sep 30 14:11:47 crc kubenswrapper[4840]: I0930 14:11:47.852961 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Sep 30 14:11:47 crc kubenswrapper[4840]: I0930 14:11:47.861374 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-txwn9"] Sep 30 14:11:47 crc kubenswrapper[4840]: I0930 14:11:47.868367 4840 generic.go:334] "Generic (PLEG): container finished" podID="377cf7b4-32f7-4aa9-85f2-1b62763d9562" containerID="14af39bb6486cffb0945112230c5f5f02c6f71930e2f48ebc53ca1dc8c694d8a" exitCode=0 Sep 30 14:11:47 crc kubenswrapper[4840]: I0930 14:11:47.868411 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-2mqrn" event={"ID":"377cf7b4-32f7-4aa9-85f2-1b62763d9562","Type":"ContainerDied","Data":"14af39bb6486cffb0945112230c5f5f02c6f71930e2f48ebc53ca1dc8c694d8a"} Sep 30 14:11:47 crc kubenswrapper[4840]: I0930 14:11:47.868485 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-2mqrn" event={"ID":"377cf7b4-32f7-4aa9-85f2-1b62763d9562","Type":"ContainerStarted","Data":"93adcd8b5e6c5f908293036f453c8596b3a9519a62b6fe0fa82a852d7d8d25b1"} Sep 30 14:11:47 crc kubenswrapper[4840]: I0930 14:11:47.933783 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d8412105-132d-4656-8fd8-296cbd8a1ace-scripts\") pod \"swift-ring-rebalance-txwn9\" (UID: \"d8412105-132d-4656-8fd8-296cbd8a1ace\") " pod="openstack/swift-ring-rebalance-txwn9" Sep 30 14:11:47 crc kubenswrapper[4840]: I0930 14:11:47.933827 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8412105-132d-4656-8fd8-296cbd8a1ace-combined-ca-bundle\") pod \"swift-ring-rebalance-txwn9\" (UID: \"d8412105-132d-4656-8fd8-296cbd8a1ace\") " pod="openstack/swift-ring-rebalance-txwn9" Sep 30 14:11:47 crc kubenswrapper[4840]: I0930 14:11:47.933890 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cpbv\" (UniqueName: \"kubernetes.io/projected/d8412105-132d-4656-8fd8-296cbd8a1ace-kube-api-access-6cpbv\") pod \"swift-ring-rebalance-txwn9\" (UID: \"d8412105-132d-4656-8fd8-296cbd8a1ace\") " pod="openstack/swift-ring-rebalance-txwn9" Sep 30 14:11:47 crc kubenswrapper[4840]: I0930 14:11:47.933907 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d8412105-132d-4656-8fd8-296cbd8a1ace-ring-data-devices\") pod \"swift-ring-rebalance-txwn9\" (UID: \"d8412105-132d-4656-8fd8-296cbd8a1ace\") " pod="openstack/swift-ring-rebalance-txwn9" Sep 30 14:11:47 crc kubenswrapper[4840]: I0930 14:11:47.933929 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d8412105-132d-4656-8fd8-296cbd8a1ace-dispersionconf\") pod \"swift-ring-rebalance-txwn9\" (UID: \"d8412105-132d-4656-8fd8-296cbd8a1ace\") " pod="openstack/swift-ring-rebalance-txwn9" Sep 30 14:11:47 crc kubenswrapper[4840]: I0930 14:11:47.934309 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d8412105-132d-4656-8fd8-296cbd8a1ace-swiftconf\") pod \"swift-ring-rebalance-txwn9\" (UID: \"d8412105-132d-4656-8fd8-296cbd8a1ace\") " pod="openstack/swift-ring-rebalance-txwn9" Sep 30 14:11:47 crc kubenswrapper[4840]: I0930 14:11:47.934471 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d8412105-132d-4656-8fd8-296cbd8a1ace-etc-swift\") pod \"swift-ring-rebalance-txwn9\" (UID: \"d8412105-132d-4656-8fd8-296cbd8a1ace\") " pod="openstack/swift-ring-rebalance-txwn9" Sep 30 14:11:48 crc kubenswrapper[4840]: I0930 14:11:48.036055 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8412105-132d-4656-8fd8-296cbd8a1ace-combined-ca-bundle\") pod \"swift-ring-rebalance-txwn9\" (UID: \"d8412105-132d-4656-8fd8-296cbd8a1ace\") " pod="openstack/swift-ring-rebalance-txwn9" Sep 30 14:11:48 crc kubenswrapper[4840]: I0930 14:11:48.036147 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/66a7a4e6-dd49-4589-bd88-2f5fdf0af70f-etc-swift\") pod \"swift-storage-0\" (UID: \"66a7a4e6-dd49-4589-bd88-2f5fdf0af70f\") " pod="openstack/swift-storage-0" Sep 30 14:11:48 crc kubenswrapper[4840]: I0930 14:11:48.036195 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cpbv\" (UniqueName: \"kubernetes.io/projected/d8412105-132d-4656-8fd8-296cbd8a1ace-kube-api-access-6cpbv\") pod \"swift-ring-rebalance-txwn9\" (UID: \"d8412105-132d-4656-8fd8-296cbd8a1ace\") " pod="openstack/swift-ring-rebalance-txwn9" Sep 30 14:11:48 crc kubenswrapper[4840]: I0930 14:11:48.036219 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d8412105-132d-4656-8fd8-296cbd8a1ace-ring-data-devices\") pod \"swift-ring-rebalance-txwn9\" (UID: \"d8412105-132d-4656-8fd8-296cbd8a1ace\") " pod="openstack/swift-ring-rebalance-txwn9" Sep 30 14:11:48 crc kubenswrapper[4840]: I0930 14:11:48.036267 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d8412105-132d-4656-8fd8-296cbd8a1ace-dispersionconf\") pod \"swift-ring-rebalance-txwn9\" (UID: \"d8412105-132d-4656-8fd8-296cbd8a1ace\") " pod="openstack/swift-ring-rebalance-txwn9" Sep 30 14:11:48 crc kubenswrapper[4840]: I0930 14:11:48.036310 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d8412105-132d-4656-8fd8-296cbd8a1ace-swiftconf\") pod \"swift-ring-rebalance-txwn9\" (UID: \"d8412105-132d-4656-8fd8-296cbd8a1ace\") " pod="openstack/swift-ring-rebalance-txwn9" Sep 30 14:11:48 crc kubenswrapper[4840]: I0930 14:11:48.036366 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d8412105-132d-4656-8fd8-296cbd8a1ace-etc-swift\") pod \"swift-ring-rebalance-txwn9\" (UID: \"d8412105-132d-4656-8fd8-296cbd8a1ace\") " pod="openstack/swift-ring-rebalance-txwn9" Sep 30 14:11:48 crc kubenswrapper[4840]: I0930 14:11:48.036509 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d8412105-132d-4656-8fd8-296cbd8a1ace-scripts\") pod \"swift-ring-rebalance-txwn9\" (UID: \"d8412105-132d-4656-8fd8-296cbd8a1ace\") " pod="openstack/swift-ring-rebalance-txwn9" Sep 30 14:11:48 crc kubenswrapper[4840]: I0930 14:11:48.037061 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d8412105-132d-4656-8fd8-296cbd8a1ace-ring-data-devices\") pod \"swift-ring-rebalance-txwn9\" (UID: \"d8412105-132d-4656-8fd8-296cbd8a1ace\") " pod="openstack/swift-ring-rebalance-txwn9" Sep 30 14:11:48 crc kubenswrapper[4840]: E0930 14:11:48.036381 4840 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Sep 30 14:11:48 crc kubenswrapper[4840]: E0930 14:11:48.037106 4840 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Sep 30 14:11:48 crc kubenswrapper[4840]: E0930 14:11:48.037140 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/66a7a4e6-dd49-4589-bd88-2f5fdf0af70f-etc-swift podName:66a7a4e6-dd49-4589-bd88-2f5fdf0af70f nodeName:}" failed. No retries permitted until 2025-09-30 14:11:49.037127179 +0000 UTC m=+937.666213592 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/66a7a4e6-dd49-4589-bd88-2f5fdf0af70f-etc-swift") pod "swift-storage-0" (UID: "66a7a4e6-dd49-4589-bd88-2f5fdf0af70f") : configmap "swift-ring-files" not found Sep 30 14:11:48 crc kubenswrapper[4840]: I0930 14:11:48.037327 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d8412105-132d-4656-8fd8-296cbd8a1ace-scripts\") pod \"swift-ring-rebalance-txwn9\" (UID: \"d8412105-132d-4656-8fd8-296cbd8a1ace\") " pod="openstack/swift-ring-rebalance-txwn9" Sep 30 14:11:48 crc kubenswrapper[4840]: I0930 14:11:48.037434 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d8412105-132d-4656-8fd8-296cbd8a1ace-etc-swift\") pod \"swift-ring-rebalance-txwn9\" (UID: \"d8412105-132d-4656-8fd8-296cbd8a1ace\") " pod="openstack/swift-ring-rebalance-txwn9" Sep 30 14:11:48 crc kubenswrapper[4840]: I0930 14:11:48.046535 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8412105-132d-4656-8fd8-296cbd8a1ace-combined-ca-bundle\") pod \"swift-ring-rebalance-txwn9\" (UID: \"d8412105-132d-4656-8fd8-296cbd8a1ace\") " pod="openstack/swift-ring-rebalance-txwn9" Sep 30 14:11:48 crc kubenswrapper[4840]: I0930 14:11:48.046907 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d8412105-132d-4656-8fd8-296cbd8a1ace-dispersionconf\") pod \"swift-ring-rebalance-txwn9\" (UID: \"d8412105-132d-4656-8fd8-296cbd8a1ace\") " pod="openstack/swift-ring-rebalance-txwn9" Sep 30 14:11:48 crc kubenswrapper[4840]: I0930 14:11:48.047477 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d8412105-132d-4656-8fd8-296cbd8a1ace-swiftconf\") pod \"swift-ring-rebalance-txwn9\" (UID: \"d8412105-132d-4656-8fd8-296cbd8a1ace\") " pod="openstack/swift-ring-rebalance-txwn9" Sep 30 14:11:48 crc kubenswrapper[4840]: I0930 14:11:48.054893 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cpbv\" (UniqueName: \"kubernetes.io/projected/d8412105-132d-4656-8fd8-296cbd8a1ace-kube-api-access-6cpbv\") pod \"swift-ring-rebalance-txwn9\" (UID: \"d8412105-132d-4656-8fd8-296cbd8a1ace\") " pod="openstack/swift-ring-rebalance-txwn9" Sep 30 14:11:48 crc kubenswrapper[4840]: I0930 14:11:48.132352 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f2e45b7-fe3e-4e53-a735-05d843a14159" path="/var/lib/kubelet/pods/9f2e45b7-fe3e-4e53-a735-05d843a14159/volumes" Sep 30 14:11:48 crc kubenswrapper[4840]: I0930 14:11:48.267859 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-txwn9" Sep 30 14:11:48 crc kubenswrapper[4840]: I0930 14:11:48.687957 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-txwn9"] Sep 30 14:11:48 crc kubenswrapper[4840]: W0930 14:11:48.691530 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8412105_132d_4656_8fd8_296cbd8a1ace.slice/crio-506501210f21c43d3c0d61e0e9471df91db2581c5dcd124819fb55fbbb361891 WatchSource:0}: Error finding container 506501210f21c43d3c0d61e0e9471df91db2581c5dcd124819fb55fbbb361891: Status 404 returned error can't find the container with id 506501210f21c43d3c0d61e0e9471df91db2581c5dcd124819fb55fbbb361891 Sep 30 14:11:48 crc kubenswrapper[4840]: I0930 14:11:48.879179 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-2mqrn" event={"ID":"377cf7b4-32f7-4aa9-85f2-1b62763d9562","Type":"ContainerStarted","Data":"d93053581c54001e96f6510a1e439aead4da6afe6b19e59c81d8f04be6a78b52"} Sep 30 14:11:48 crc kubenswrapper[4840]: I0930 14:11:48.879348 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-2mqrn" Sep 30 14:11:48 crc kubenswrapper[4840]: I0930 14:11:48.880603 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-txwn9" event={"ID":"d8412105-132d-4656-8fd8-296cbd8a1ace","Type":"ContainerStarted","Data":"506501210f21c43d3c0d61e0e9471df91db2581c5dcd124819fb55fbbb361891"} Sep 30 14:11:48 crc kubenswrapper[4840]: I0930 14:11:48.905697 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-2mqrn" podStartSLOduration=2.9056785830000003 podStartE2EDuration="2.905678583s" podCreationTimestamp="2025-09-30 14:11:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:11:48.89940956 +0000 UTC m=+937.528495983" watchObservedRunningTime="2025-09-30 14:11:48.905678583 +0000 UTC m=+937.534765016" Sep 30 14:11:49 crc kubenswrapper[4840]: I0930 14:11:49.052503 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/66a7a4e6-dd49-4589-bd88-2f5fdf0af70f-etc-swift\") pod \"swift-storage-0\" (UID: \"66a7a4e6-dd49-4589-bd88-2f5fdf0af70f\") " pod="openstack/swift-storage-0" Sep 30 14:11:49 crc kubenswrapper[4840]: E0930 14:11:49.052729 4840 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Sep 30 14:11:49 crc kubenswrapper[4840]: E0930 14:11:49.052752 4840 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Sep 30 14:11:49 crc kubenswrapper[4840]: E0930 14:11:49.052811 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/66a7a4e6-dd49-4589-bd88-2f5fdf0af70f-etc-swift podName:66a7a4e6-dd49-4589-bd88-2f5fdf0af70f nodeName:}" failed. No retries permitted until 2025-09-30 14:11:51.052791878 +0000 UTC m=+939.681878301 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/66a7a4e6-dd49-4589-bd88-2f5fdf0af70f-etc-swift") pod "swift-storage-0" (UID: "66a7a4e6-dd49-4589-bd88-2f5fdf0af70f") : configmap "swift-ring-files" not found Sep 30 14:11:49 crc kubenswrapper[4840]: I0930 14:11:49.444238 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-n7hj6"] Sep 30 14:11:49 crc kubenswrapper[4840]: I0930 14:11:49.446492 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-n7hj6" Sep 30 14:11:49 crc kubenswrapper[4840]: I0930 14:11:49.461284 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8tvg\" (UniqueName: \"kubernetes.io/projected/abd36ef0-abb9-4812-adcb-f3b56172b81e-kube-api-access-l8tvg\") pod \"glance-db-create-n7hj6\" (UID: \"abd36ef0-abb9-4812-adcb-f3b56172b81e\") " pod="openstack/glance-db-create-n7hj6" Sep 30 14:11:49 crc kubenswrapper[4840]: I0930 14:11:49.480149 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-n7hj6"] Sep 30 14:11:49 crc kubenswrapper[4840]: I0930 14:11:49.562584 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8tvg\" (UniqueName: \"kubernetes.io/projected/abd36ef0-abb9-4812-adcb-f3b56172b81e-kube-api-access-l8tvg\") pod \"glance-db-create-n7hj6\" (UID: \"abd36ef0-abb9-4812-adcb-f3b56172b81e\") " pod="openstack/glance-db-create-n7hj6" Sep 30 14:11:49 crc kubenswrapper[4840]: I0930 14:11:49.582163 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8tvg\" (UniqueName: \"kubernetes.io/projected/abd36ef0-abb9-4812-adcb-f3b56172b81e-kube-api-access-l8tvg\") pod \"glance-db-create-n7hj6\" (UID: \"abd36ef0-abb9-4812-adcb-f3b56172b81e\") " pod="openstack/glance-db-create-n7hj6" Sep 30 14:11:49 crc kubenswrapper[4840]: I0930 14:11:49.781207 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-n7hj6" Sep 30 14:11:50 crc kubenswrapper[4840]: I0930 14:11:50.236878 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-n7hj6"] Sep 30 14:11:50 crc kubenswrapper[4840]: W0930 14:11:50.244441 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podabd36ef0_abb9_4812_adcb_f3b56172b81e.slice/crio-791df1e053339a0ed9378caf8f554f5f5419dc8b41ed98f3db6b1ad72bb30292 WatchSource:0}: Error finding container 791df1e053339a0ed9378caf8f554f5f5419dc8b41ed98f3db6b1ad72bb30292: Status 404 returned error can't find the container with id 791df1e053339a0ed9378caf8f554f5f5419dc8b41ed98f3db6b1ad72bb30292 Sep 30 14:11:50 crc kubenswrapper[4840]: I0930 14:11:50.906353 4840 generic.go:334] "Generic (PLEG): container finished" podID="abd36ef0-abb9-4812-adcb-f3b56172b81e" containerID="2d128c0e805ecf2486bb4310324e54729f86c76e69942a710a8ba7dafedf57d5" exitCode=0 Sep 30 14:11:50 crc kubenswrapper[4840]: I0930 14:11:50.906392 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-n7hj6" event={"ID":"abd36ef0-abb9-4812-adcb-f3b56172b81e","Type":"ContainerDied","Data":"2d128c0e805ecf2486bb4310324e54729f86c76e69942a710a8ba7dafedf57d5"} Sep 30 14:11:50 crc kubenswrapper[4840]: I0930 14:11:50.906699 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-n7hj6" event={"ID":"abd36ef0-abb9-4812-adcb-f3b56172b81e","Type":"ContainerStarted","Data":"791df1e053339a0ed9378caf8f554f5f5419dc8b41ed98f3db6b1ad72bb30292"} Sep 30 14:11:51 crc kubenswrapper[4840]: I0930 14:11:51.087986 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/66a7a4e6-dd49-4589-bd88-2f5fdf0af70f-etc-swift\") pod \"swift-storage-0\" (UID: \"66a7a4e6-dd49-4589-bd88-2f5fdf0af70f\") " pod="openstack/swift-storage-0" Sep 30 14:11:51 crc kubenswrapper[4840]: E0930 14:11:51.088174 4840 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Sep 30 14:11:51 crc kubenswrapper[4840]: E0930 14:11:51.088233 4840 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Sep 30 14:11:51 crc kubenswrapper[4840]: E0930 14:11:51.088317 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/66a7a4e6-dd49-4589-bd88-2f5fdf0af70f-etc-swift podName:66a7a4e6-dd49-4589-bd88-2f5fdf0af70f nodeName:}" failed. No retries permitted until 2025-09-30 14:11:55.088289174 +0000 UTC m=+943.717375637 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/66a7a4e6-dd49-4589-bd88-2f5fdf0af70f-etc-swift") pod "swift-storage-0" (UID: "66a7a4e6-dd49-4589-bd88-2f5fdf0af70f") : configmap "swift-ring-files" not found Sep 30 14:11:52 crc kubenswrapper[4840]: I0930 14:11:52.257940 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-n7hj6" Sep 30 14:11:52 crc kubenswrapper[4840]: I0930 14:11:52.308188 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8tvg\" (UniqueName: \"kubernetes.io/projected/abd36ef0-abb9-4812-adcb-f3b56172b81e-kube-api-access-l8tvg\") pod \"abd36ef0-abb9-4812-adcb-f3b56172b81e\" (UID: \"abd36ef0-abb9-4812-adcb-f3b56172b81e\") " Sep 30 14:11:52 crc kubenswrapper[4840]: I0930 14:11:52.314537 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abd36ef0-abb9-4812-adcb-f3b56172b81e-kube-api-access-l8tvg" (OuterVolumeSpecName: "kube-api-access-l8tvg") pod "abd36ef0-abb9-4812-adcb-f3b56172b81e" (UID: "abd36ef0-abb9-4812-adcb-f3b56172b81e"). InnerVolumeSpecName "kube-api-access-l8tvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:11:52 crc kubenswrapper[4840]: I0930 14:11:52.410009 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8tvg\" (UniqueName: \"kubernetes.io/projected/abd36ef0-abb9-4812-adcb-f3b56172b81e-kube-api-access-l8tvg\") on node \"crc\" DevicePath \"\"" Sep 30 14:11:52 crc kubenswrapper[4840]: I0930 14:11:52.928392 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-n7hj6" Sep 30 14:11:52 crc kubenswrapper[4840]: I0930 14:11:52.928404 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-n7hj6" event={"ID":"abd36ef0-abb9-4812-adcb-f3b56172b81e","Type":"ContainerDied","Data":"791df1e053339a0ed9378caf8f554f5f5419dc8b41ed98f3db6b1ad72bb30292"} Sep 30 14:11:52 crc kubenswrapper[4840]: I0930 14:11:52.928897 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="791df1e053339a0ed9378caf8f554f5f5419dc8b41ed98f3db6b1ad72bb30292" Sep 30 14:11:52 crc kubenswrapper[4840]: I0930 14:11:52.931011 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-txwn9" event={"ID":"d8412105-132d-4656-8fd8-296cbd8a1ace","Type":"ContainerStarted","Data":"5bada32f93188343f24b507853e104d6725fdd3f35b081b217e996756f0db1d7"} Sep 30 14:11:52 crc kubenswrapper[4840]: I0930 14:11:52.965944 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-txwn9" podStartSLOduration=2.544967326 podStartE2EDuration="5.965915066s" podCreationTimestamp="2025-09-30 14:11:47 +0000 UTC" firstStartedPulling="2025-09-30 14:11:48.693894356 +0000 UTC m=+937.322980779" lastFinishedPulling="2025-09-30 14:11:52.114842096 +0000 UTC m=+940.743928519" observedRunningTime="2025-09-30 14:11:52.951892021 +0000 UTC m=+941.580978484" watchObservedRunningTime="2025-09-30 14:11:52.965915066 +0000 UTC m=+941.595001509" Sep 30 14:11:53 crc kubenswrapper[4840]: I0930 14:11:53.330429 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Sep 30 14:11:53 crc kubenswrapper[4840]: I0930 14:11:53.750173 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-g7vsc"] Sep 30 14:11:53 crc kubenswrapper[4840]: E0930 14:11:53.751122 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abd36ef0-abb9-4812-adcb-f3b56172b81e" containerName="mariadb-database-create" Sep 30 14:11:53 crc kubenswrapper[4840]: I0930 14:11:53.751217 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="abd36ef0-abb9-4812-adcb-f3b56172b81e" containerName="mariadb-database-create" Sep 30 14:11:53 crc kubenswrapper[4840]: I0930 14:11:53.751652 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="abd36ef0-abb9-4812-adcb-f3b56172b81e" containerName="mariadb-database-create" Sep 30 14:11:53 crc kubenswrapper[4840]: I0930 14:11:53.752349 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-g7vsc" Sep 30 14:11:53 crc kubenswrapper[4840]: I0930 14:11:53.756398 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-g7vsc"] Sep 30 14:11:53 crc kubenswrapper[4840]: I0930 14:11:53.829359 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsx5c\" (UniqueName: \"kubernetes.io/projected/46aeb27d-7764-4fbb-bbdb-88628a7d6f59-kube-api-access-vsx5c\") pod \"keystone-db-create-g7vsc\" (UID: \"46aeb27d-7764-4fbb-bbdb-88628a7d6f59\") " pod="openstack/keystone-db-create-g7vsc" Sep 30 14:11:53 crc kubenswrapper[4840]: I0930 14:11:53.930494 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsx5c\" (UniqueName: \"kubernetes.io/projected/46aeb27d-7764-4fbb-bbdb-88628a7d6f59-kube-api-access-vsx5c\") pod \"keystone-db-create-g7vsc\" (UID: \"46aeb27d-7764-4fbb-bbdb-88628a7d6f59\") " pod="openstack/keystone-db-create-g7vsc" Sep 30 14:11:53 crc kubenswrapper[4840]: I0930 14:11:53.955331 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsx5c\" (UniqueName: \"kubernetes.io/projected/46aeb27d-7764-4fbb-bbdb-88628a7d6f59-kube-api-access-vsx5c\") pod \"keystone-db-create-g7vsc\" (UID: \"46aeb27d-7764-4fbb-bbdb-88628a7d6f59\") " pod="openstack/keystone-db-create-g7vsc" Sep 30 14:11:54 crc kubenswrapper[4840]: I0930 14:11:54.061005 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-5mfh2"] Sep 30 14:11:54 crc kubenswrapper[4840]: I0930 14:11:54.062370 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5mfh2" Sep 30 14:11:54 crc kubenswrapper[4840]: I0930 14:11:54.068466 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-g7vsc" Sep 30 14:11:54 crc kubenswrapper[4840]: I0930 14:11:54.069386 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-5mfh2"] Sep 30 14:11:54 crc kubenswrapper[4840]: I0930 14:11:54.133634 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clpht\" (UniqueName: \"kubernetes.io/projected/9bb72efc-899f-4e40-a519-86402ed702b7-kube-api-access-clpht\") pod \"placement-db-create-5mfh2\" (UID: \"9bb72efc-899f-4e40-a519-86402ed702b7\") " pod="openstack/placement-db-create-5mfh2" Sep 30 14:11:54 crc kubenswrapper[4840]: I0930 14:11:54.236325 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clpht\" (UniqueName: \"kubernetes.io/projected/9bb72efc-899f-4e40-a519-86402ed702b7-kube-api-access-clpht\") pod \"placement-db-create-5mfh2\" (UID: \"9bb72efc-899f-4e40-a519-86402ed702b7\") " pod="openstack/placement-db-create-5mfh2" Sep 30 14:11:54 crc kubenswrapper[4840]: I0930 14:11:54.263731 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clpht\" (UniqueName: \"kubernetes.io/projected/9bb72efc-899f-4e40-a519-86402ed702b7-kube-api-access-clpht\") pod \"placement-db-create-5mfh2\" (UID: \"9bb72efc-899f-4e40-a519-86402ed702b7\") " pod="openstack/placement-db-create-5mfh2" Sep 30 14:11:54 crc kubenswrapper[4840]: I0930 14:11:54.382273 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5mfh2" Sep 30 14:11:54 crc kubenswrapper[4840]: I0930 14:11:54.493984 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-g7vsc"] Sep 30 14:11:54 crc kubenswrapper[4840]: I0930 14:11:54.807083 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-5mfh2"] Sep 30 14:11:54 crc kubenswrapper[4840]: I0930 14:11:54.944476 4840 generic.go:334] "Generic (PLEG): container finished" podID="46aeb27d-7764-4fbb-bbdb-88628a7d6f59" containerID="26e779802fe5efa670adadddea2105ace90b73864327733e492daf4ef27f174d" exitCode=0 Sep 30 14:11:54 crc kubenswrapper[4840]: I0930 14:11:54.944570 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-g7vsc" event={"ID":"46aeb27d-7764-4fbb-bbdb-88628a7d6f59","Type":"ContainerDied","Data":"26e779802fe5efa670adadddea2105ace90b73864327733e492daf4ef27f174d"} Sep 30 14:11:54 crc kubenswrapper[4840]: I0930 14:11:54.944596 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-g7vsc" event={"ID":"46aeb27d-7764-4fbb-bbdb-88628a7d6f59","Type":"ContainerStarted","Data":"bcea7743913a73ede7b91ef50c19eb42ebd5a0e693b5e2169b6a18e7189fd413"} Sep 30 14:11:54 crc kubenswrapper[4840]: I0930 14:11:54.946007 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-5mfh2" event={"ID":"9bb72efc-899f-4e40-a519-86402ed702b7","Type":"ContainerStarted","Data":"55ce9a2b8c0bde8dcaaa1589f64cea57c71d079ce974061d3d2a6b9e5fae8594"} Sep 30 14:11:54 crc kubenswrapper[4840]: I0930 14:11:54.946030 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-5mfh2" event={"ID":"9bb72efc-899f-4e40-a519-86402ed702b7","Type":"ContainerStarted","Data":"97f3ca6ad8243ab279d3fed5342d3c915c57fe4cd5505a2c80b09eb7b921e52d"} Sep 30 14:11:54 crc kubenswrapper[4840]: I0930 14:11:54.977117 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-5mfh2" podStartSLOduration=0.977093669 podStartE2EDuration="977.093669ms" podCreationTimestamp="2025-09-30 14:11:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:11:54.972222182 +0000 UTC m=+943.601308645" watchObservedRunningTime="2025-09-30 14:11:54.977093669 +0000 UTC m=+943.606180102" Sep 30 14:11:55 crc kubenswrapper[4840]: I0930 14:11:55.163421 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/66a7a4e6-dd49-4589-bd88-2f5fdf0af70f-etc-swift\") pod \"swift-storage-0\" (UID: \"66a7a4e6-dd49-4589-bd88-2f5fdf0af70f\") " pod="openstack/swift-storage-0" Sep 30 14:11:55 crc kubenswrapper[4840]: E0930 14:11:55.163481 4840 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Sep 30 14:11:55 crc kubenswrapper[4840]: E0930 14:11:55.163501 4840 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Sep 30 14:11:55 crc kubenswrapper[4840]: E0930 14:11:55.163686 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/66a7a4e6-dd49-4589-bd88-2f5fdf0af70f-etc-swift podName:66a7a4e6-dd49-4589-bd88-2f5fdf0af70f nodeName:}" failed. No retries permitted until 2025-09-30 14:12:03.16366103 +0000 UTC m=+951.792747443 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/66a7a4e6-dd49-4589-bd88-2f5fdf0af70f-etc-swift") pod "swift-storage-0" (UID: "66a7a4e6-dd49-4589-bd88-2f5fdf0af70f") : configmap "swift-ring-files" not found Sep 30 14:11:55 crc kubenswrapper[4840]: I0930 14:11:55.961804 4840 generic.go:334] "Generic (PLEG): container finished" podID="9bb72efc-899f-4e40-a519-86402ed702b7" containerID="55ce9a2b8c0bde8dcaaa1589f64cea57c71d079ce974061d3d2a6b9e5fae8594" exitCode=0 Sep 30 14:11:55 crc kubenswrapper[4840]: I0930 14:11:55.962287 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-5mfh2" event={"ID":"9bb72efc-899f-4e40-a519-86402ed702b7","Type":"ContainerDied","Data":"55ce9a2b8c0bde8dcaaa1589f64cea57c71d079ce974061d3d2a6b9e5fae8594"} Sep 30 14:11:56 crc kubenswrapper[4840]: I0930 14:11:56.265282 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-g7vsc" Sep 30 14:11:56 crc kubenswrapper[4840]: I0930 14:11:56.384469 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vsx5c\" (UniqueName: \"kubernetes.io/projected/46aeb27d-7764-4fbb-bbdb-88628a7d6f59-kube-api-access-vsx5c\") pod \"46aeb27d-7764-4fbb-bbdb-88628a7d6f59\" (UID: \"46aeb27d-7764-4fbb-bbdb-88628a7d6f59\") " Sep 30 14:11:56 crc kubenswrapper[4840]: I0930 14:11:56.390359 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46aeb27d-7764-4fbb-bbdb-88628a7d6f59-kube-api-access-vsx5c" (OuterVolumeSpecName: "kube-api-access-vsx5c") pod "46aeb27d-7764-4fbb-bbdb-88628a7d6f59" (UID: "46aeb27d-7764-4fbb-bbdb-88628a7d6f59"). InnerVolumeSpecName "kube-api-access-vsx5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:11:56 crc kubenswrapper[4840]: I0930 14:11:56.487215 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vsx5c\" (UniqueName: \"kubernetes.io/projected/46aeb27d-7764-4fbb-bbdb-88628a7d6f59-kube-api-access-vsx5c\") on node \"crc\" DevicePath \"\"" Sep 30 14:11:56 crc kubenswrapper[4840]: I0930 14:11:56.531741 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-2mqrn" Sep 30 14:11:56 crc kubenswrapper[4840]: I0930 14:11:56.602041 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-gt8xx"] Sep 30 14:11:56 crc kubenswrapper[4840]: I0930 14:11:56.602314 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-gt8xx" podUID="f2759eed-4989-43b4-b025-4aab60dc864a" containerName="dnsmasq-dns" containerID="cri-o://81d2e9ea2e06c3dd73796140bb1862f4d42eeb2f12f035a1619c6007f2483bb7" gracePeriod=10 Sep 30 14:11:56 crc kubenswrapper[4840]: I0930 14:11:56.970813 4840 generic.go:334] "Generic (PLEG): container finished" podID="f2759eed-4989-43b4-b025-4aab60dc864a" containerID="81d2e9ea2e06c3dd73796140bb1862f4d42eeb2f12f035a1619c6007f2483bb7" exitCode=0 Sep 30 14:11:56 crc kubenswrapper[4840]: I0930 14:11:56.970876 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-gt8xx" event={"ID":"f2759eed-4989-43b4-b025-4aab60dc864a","Type":"ContainerDied","Data":"81d2e9ea2e06c3dd73796140bb1862f4d42eeb2f12f035a1619c6007f2483bb7"} Sep 30 14:11:56 crc kubenswrapper[4840]: I0930 14:11:56.970903 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-gt8xx" event={"ID":"f2759eed-4989-43b4-b025-4aab60dc864a","Type":"ContainerDied","Data":"05bf60d61c0aca11f948fba1e90455fd5be9e1c7f953f6b9f16bbbf0551dab8d"} Sep 30 14:11:56 crc kubenswrapper[4840]: I0930 14:11:56.970915 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="05bf60d61c0aca11f948fba1e90455fd5be9e1c7f953f6b9f16bbbf0551dab8d" Sep 30 14:11:56 crc kubenswrapper[4840]: I0930 14:11:56.972349 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-g7vsc" Sep 30 14:11:56 crc kubenswrapper[4840]: I0930 14:11:56.972396 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-g7vsc" event={"ID":"46aeb27d-7764-4fbb-bbdb-88628a7d6f59","Type":"ContainerDied","Data":"bcea7743913a73ede7b91ef50c19eb42ebd5a0e693b5e2169b6a18e7189fd413"} Sep 30 14:11:56 crc kubenswrapper[4840]: I0930 14:11:56.972414 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bcea7743913a73ede7b91ef50c19eb42ebd5a0e693b5e2169b6a18e7189fd413" Sep 30 14:11:57 crc kubenswrapper[4840]: I0930 14:11:57.011720 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-gt8xx" Sep 30 14:11:57 crc kubenswrapper[4840]: I0930 14:11:57.097298 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f2759eed-4989-43b4-b025-4aab60dc864a-dns-svc\") pod \"f2759eed-4989-43b4-b025-4aab60dc864a\" (UID: \"f2759eed-4989-43b4-b025-4aab60dc864a\") " Sep 30 14:11:57 crc kubenswrapper[4840]: I0930 14:11:57.097367 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtz4t\" (UniqueName: \"kubernetes.io/projected/f2759eed-4989-43b4-b025-4aab60dc864a-kube-api-access-mtz4t\") pod \"f2759eed-4989-43b4-b025-4aab60dc864a\" (UID: \"f2759eed-4989-43b4-b025-4aab60dc864a\") " Sep 30 14:11:57 crc kubenswrapper[4840]: I0930 14:11:57.097392 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2759eed-4989-43b4-b025-4aab60dc864a-config\") pod \"f2759eed-4989-43b4-b025-4aab60dc864a\" (UID: \"f2759eed-4989-43b4-b025-4aab60dc864a\") " Sep 30 14:11:57 crc kubenswrapper[4840]: I0930 14:11:57.097436 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f2759eed-4989-43b4-b025-4aab60dc864a-ovsdbserver-nb\") pod \"f2759eed-4989-43b4-b025-4aab60dc864a\" (UID: \"f2759eed-4989-43b4-b025-4aab60dc864a\") " Sep 30 14:11:57 crc kubenswrapper[4840]: I0930 14:11:57.097481 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f2759eed-4989-43b4-b025-4aab60dc864a-ovsdbserver-sb\") pod \"f2759eed-4989-43b4-b025-4aab60dc864a\" (UID: \"f2759eed-4989-43b4-b025-4aab60dc864a\") " Sep 30 14:11:57 crc kubenswrapper[4840]: I0930 14:11:57.101506 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2759eed-4989-43b4-b025-4aab60dc864a-kube-api-access-mtz4t" (OuterVolumeSpecName: "kube-api-access-mtz4t") pod "f2759eed-4989-43b4-b025-4aab60dc864a" (UID: "f2759eed-4989-43b4-b025-4aab60dc864a"). InnerVolumeSpecName "kube-api-access-mtz4t". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:11:57 crc kubenswrapper[4840]: I0930 14:11:57.145376 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2759eed-4989-43b4-b025-4aab60dc864a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f2759eed-4989-43b4-b025-4aab60dc864a" (UID: "f2759eed-4989-43b4-b025-4aab60dc864a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:11:57 crc kubenswrapper[4840]: I0930 14:11:57.155154 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2759eed-4989-43b4-b025-4aab60dc864a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f2759eed-4989-43b4-b025-4aab60dc864a" (UID: "f2759eed-4989-43b4-b025-4aab60dc864a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:11:57 crc kubenswrapper[4840]: I0930 14:11:57.160703 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2759eed-4989-43b4-b025-4aab60dc864a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f2759eed-4989-43b4-b025-4aab60dc864a" (UID: "f2759eed-4989-43b4-b025-4aab60dc864a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:11:57 crc kubenswrapper[4840]: I0930 14:11:57.162999 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2759eed-4989-43b4-b025-4aab60dc864a-config" (OuterVolumeSpecName: "config") pod "f2759eed-4989-43b4-b025-4aab60dc864a" (UID: "f2759eed-4989-43b4-b025-4aab60dc864a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:11:57 crc kubenswrapper[4840]: I0930 14:11:57.199378 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtz4t\" (UniqueName: \"kubernetes.io/projected/f2759eed-4989-43b4-b025-4aab60dc864a-kube-api-access-mtz4t\") on node \"crc\" DevicePath \"\"" Sep 30 14:11:57 crc kubenswrapper[4840]: I0930 14:11:57.199620 4840 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2759eed-4989-43b4-b025-4aab60dc864a-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:11:57 crc kubenswrapper[4840]: I0930 14:11:57.199697 4840 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f2759eed-4989-43b4-b025-4aab60dc864a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 14:11:57 crc kubenswrapper[4840]: I0930 14:11:57.199786 4840 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f2759eed-4989-43b4-b025-4aab60dc864a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 14:11:57 crc kubenswrapper[4840]: I0930 14:11:57.199851 4840 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f2759eed-4989-43b4-b025-4aab60dc864a-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 14:11:57 crc kubenswrapper[4840]: I0930 14:11:57.283833 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5mfh2" Sep 30 14:11:57 crc kubenswrapper[4840]: I0930 14:11:57.402856 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-clpht\" (UniqueName: \"kubernetes.io/projected/9bb72efc-899f-4e40-a519-86402ed702b7-kube-api-access-clpht\") pod \"9bb72efc-899f-4e40-a519-86402ed702b7\" (UID: \"9bb72efc-899f-4e40-a519-86402ed702b7\") " Sep 30 14:11:57 crc kubenswrapper[4840]: I0930 14:11:57.406638 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9bb72efc-899f-4e40-a519-86402ed702b7-kube-api-access-clpht" (OuterVolumeSpecName: "kube-api-access-clpht") pod "9bb72efc-899f-4e40-a519-86402ed702b7" (UID: "9bb72efc-899f-4e40-a519-86402ed702b7"). InnerVolumeSpecName "kube-api-access-clpht". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:11:57 crc kubenswrapper[4840]: I0930 14:11:57.504581 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-clpht\" (UniqueName: \"kubernetes.io/projected/9bb72efc-899f-4e40-a519-86402ed702b7-kube-api-access-clpht\") on node \"crc\" DevicePath \"\"" Sep 30 14:11:57 crc kubenswrapper[4840]: I0930 14:11:57.982974 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-gt8xx" Sep 30 14:11:57 crc kubenswrapper[4840]: I0930 14:11:57.983005 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5mfh2" Sep 30 14:11:57 crc kubenswrapper[4840]: I0930 14:11:57.983066 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-5mfh2" event={"ID":"9bb72efc-899f-4e40-a519-86402ed702b7","Type":"ContainerDied","Data":"97f3ca6ad8243ab279d3fed5342d3c915c57fe4cd5505a2c80b09eb7b921e52d"} Sep 30 14:11:57 crc kubenswrapper[4840]: I0930 14:11:57.983101 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="97f3ca6ad8243ab279d3fed5342d3c915c57fe4cd5505a2c80b09eb7b921e52d" Sep 30 14:11:58 crc kubenswrapper[4840]: I0930 14:11:58.027906 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-gt8xx"] Sep 30 14:11:58 crc kubenswrapper[4840]: I0930 14:11:58.038646 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-gt8xx"] Sep 30 14:11:58 crc kubenswrapper[4840]: I0930 14:11:58.132322 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2759eed-4989-43b4-b025-4aab60dc864a" path="/var/lib/kubelet/pods/f2759eed-4989-43b4-b025-4aab60dc864a/volumes" Sep 30 14:11:58 crc kubenswrapper[4840]: I0930 14:11:58.995603 4840 generic.go:334] "Generic (PLEG): container finished" podID="d8412105-132d-4656-8fd8-296cbd8a1ace" containerID="5bada32f93188343f24b507853e104d6725fdd3f35b081b217e996756f0db1d7" exitCode=0 Sep 30 14:11:58 crc kubenswrapper[4840]: I0930 14:11:58.995650 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-txwn9" event={"ID":"d8412105-132d-4656-8fd8-296cbd8a1ace","Type":"ContainerDied","Data":"5bada32f93188343f24b507853e104d6725fdd3f35b081b217e996756f0db1d7"} Sep 30 14:11:59 crc kubenswrapper[4840]: I0930 14:11:59.538862 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-ae90-account-create-4schj"] Sep 30 14:11:59 crc kubenswrapper[4840]: E0930 14:11:59.539651 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2759eed-4989-43b4-b025-4aab60dc864a" containerName="init" Sep 30 14:11:59 crc kubenswrapper[4840]: I0930 14:11:59.539676 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2759eed-4989-43b4-b025-4aab60dc864a" containerName="init" Sep 30 14:11:59 crc kubenswrapper[4840]: E0930 14:11:59.539699 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2759eed-4989-43b4-b025-4aab60dc864a" containerName="dnsmasq-dns" Sep 30 14:11:59 crc kubenswrapper[4840]: I0930 14:11:59.539711 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2759eed-4989-43b4-b025-4aab60dc864a" containerName="dnsmasq-dns" Sep 30 14:11:59 crc kubenswrapper[4840]: E0930 14:11:59.539772 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bb72efc-899f-4e40-a519-86402ed702b7" containerName="mariadb-database-create" Sep 30 14:11:59 crc kubenswrapper[4840]: I0930 14:11:59.539784 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bb72efc-899f-4e40-a519-86402ed702b7" containerName="mariadb-database-create" Sep 30 14:11:59 crc kubenswrapper[4840]: E0930 14:11:59.539797 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46aeb27d-7764-4fbb-bbdb-88628a7d6f59" containerName="mariadb-database-create" Sep 30 14:11:59 crc kubenswrapper[4840]: I0930 14:11:59.539809 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="46aeb27d-7764-4fbb-bbdb-88628a7d6f59" containerName="mariadb-database-create" Sep 30 14:11:59 crc kubenswrapper[4840]: I0930 14:11:59.540068 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="46aeb27d-7764-4fbb-bbdb-88628a7d6f59" containerName="mariadb-database-create" Sep 30 14:11:59 crc kubenswrapper[4840]: I0930 14:11:59.540087 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="9bb72efc-899f-4e40-a519-86402ed702b7" containerName="mariadb-database-create" Sep 30 14:11:59 crc kubenswrapper[4840]: I0930 14:11:59.540101 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2759eed-4989-43b4-b025-4aab60dc864a" containerName="dnsmasq-dns" Sep 30 14:11:59 crc kubenswrapper[4840]: I0930 14:11:59.540908 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ae90-account-create-4schj" Sep 30 14:11:59 crc kubenswrapper[4840]: I0930 14:11:59.545979 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Sep 30 14:11:59 crc kubenswrapper[4840]: I0930 14:11:59.558175 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-ae90-account-create-4schj"] Sep 30 14:11:59 crc kubenswrapper[4840]: I0930 14:11:59.636618 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fn7lz\" (UniqueName: \"kubernetes.io/projected/b5f49283-d818-4180-9b52-4f05352586e4-kube-api-access-fn7lz\") pod \"glance-ae90-account-create-4schj\" (UID: \"b5f49283-d818-4180-9b52-4f05352586e4\") " pod="openstack/glance-ae90-account-create-4schj" Sep 30 14:11:59 crc kubenswrapper[4840]: I0930 14:11:59.737644 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fn7lz\" (UniqueName: \"kubernetes.io/projected/b5f49283-d818-4180-9b52-4f05352586e4-kube-api-access-fn7lz\") pod \"glance-ae90-account-create-4schj\" (UID: \"b5f49283-d818-4180-9b52-4f05352586e4\") " pod="openstack/glance-ae90-account-create-4schj" Sep 30 14:11:59 crc kubenswrapper[4840]: I0930 14:11:59.753813 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fn7lz\" (UniqueName: \"kubernetes.io/projected/b5f49283-d818-4180-9b52-4f05352586e4-kube-api-access-fn7lz\") pod \"glance-ae90-account-create-4schj\" (UID: \"b5f49283-d818-4180-9b52-4f05352586e4\") " pod="openstack/glance-ae90-account-create-4schj" Sep 30 14:11:59 crc kubenswrapper[4840]: I0930 14:11:59.882778 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ae90-account-create-4schj" Sep 30 14:12:00 crc kubenswrapper[4840]: I0930 14:12:00.348390 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-txwn9" Sep 30 14:12:00 crc kubenswrapper[4840]: I0930 14:12:00.359093 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-ae90-account-create-4schj"] Sep 30 14:12:00 crc kubenswrapper[4840]: I0930 14:12:00.549163 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d8412105-132d-4656-8fd8-296cbd8a1ace-swiftconf\") pod \"d8412105-132d-4656-8fd8-296cbd8a1ace\" (UID: \"d8412105-132d-4656-8fd8-296cbd8a1ace\") " Sep 30 14:12:00 crc kubenswrapper[4840]: I0930 14:12:00.549230 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d8412105-132d-4656-8fd8-296cbd8a1ace-scripts\") pod \"d8412105-132d-4656-8fd8-296cbd8a1ace\" (UID: \"d8412105-132d-4656-8fd8-296cbd8a1ace\") " Sep 30 14:12:00 crc kubenswrapper[4840]: I0930 14:12:00.549266 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6cpbv\" (UniqueName: \"kubernetes.io/projected/d8412105-132d-4656-8fd8-296cbd8a1ace-kube-api-access-6cpbv\") pod \"d8412105-132d-4656-8fd8-296cbd8a1ace\" (UID: \"d8412105-132d-4656-8fd8-296cbd8a1ace\") " Sep 30 14:12:00 crc kubenswrapper[4840]: I0930 14:12:00.549294 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d8412105-132d-4656-8fd8-296cbd8a1ace-dispersionconf\") pod \"d8412105-132d-4656-8fd8-296cbd8a1ace\" (UID: \"d8412105-132d-4656-8fd8-296cbd8a1ace\") " Sep 30 14:12:00 crc kubenswrapper[4840]: I0930 14:12:00.549318 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d8412105-132d-4656-8fd8-296cbd8a1ace-ring-data-devices\") pod \"d8412105-132d-4656-8fd8-296cbd8a1ace\" (UID: \"d8412105-132d-4656-8fd8-296cbd8a1ace\") " Sep 30 14:12:00 crc kubenswrapper[4840]: I0930 14:12:00.549348 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d8412105-132d-4656-8fd8-296cbd8a1ace-etc-swift\") pod \"d8412105-132d-4656-8fd8-296cbd8a1ace\" (UID: \"d8412105-132d-4656-8fd8-296cbd8a1ace\") " Sep 30 14:12:00 crc kubenswrapper[4840]: I0930 14:12:00.549402 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8412105-132d-4656-8fd8-296cbd8a1ace-combined-ca-bundle\") pod \"d8412105-132d-4656-8fd8-296cbd8a1ace\" (UID: \"d8412105-132d-4656-8fd8-296cbd8a1ace\") " Sep 30 14:12:00 crc kubenswrapper[4840]: I0930 14:12:00.551538 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8412105-132d-4656-8fd8-296cbd8a1ace-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "d8412105-132d-4656-8fd8-296cbd8a1ace" (UID: "d8412105-132d-4656-8fd8-296cbd8a1ace"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:12:00 crc kubenswrapper[4840]: I0930 14:12:00.551806 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8412105-132d-4656-8fd8-296cbd8a1ace-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "d8412105-132d-4656-8fd8-296cbd8a1ace" (UID: "d8412105-132d-4656-8fd8-296cbd8a1ace"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:12:00 crc kubenswrapper[4840]: I0930 14:12:00.555791 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8412105-132d-4656-8fd8-296cbd8a1ace-kube-api-access-6cpbv" (OuterVolumeSpecName: "kube-api-access-6cpbv") pod "d8412105-132d-4656-8fd8-296cbd8a1ace" (UID: "d8412105-132d-4656-8fd8-296cbd8a1ace"). InnerVolumeSpecName "kube-api-access-6cpbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:12:00 crc kubenswrapper[4840]: I0930 14:12:00.557899 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8412105-132d-4656-8fd8-296cbd8a1ace-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "d8412105-132d-4656-8fd8-296cbd8a1ace" (UID: "d8412105-132d-4656-8fd8-296cbd8a1ace"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:12:00 crc kubenswrapper[4840]: I0930 14:12:00.572656 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8412105-132d-4656-8fd8-296cbd8a1ace-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d8412105-132d-4656-8fd8-296cbd8a1ace" (UID: "d8412105-132d-4656-8fd8-296cbd8a1ace"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:12:00 crc kubenswrapper[4840]: I0930 14:12:00.573332 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8412105-132d-4656-8fd8-296cbd8a1ace-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "d8412105-132d-4656-8fd8-296cbd8a1ace" (UID: "d8412105-132d-4656-8fd8-296cbd8a1ace"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:12:00 crc kubenswrapper[4840]: I0930 14:12:00.585404 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8412105-132d-4656-8fd8-296cbd8a1ace-scripts" (OuterVolumeSpecName: "scripts") pod "d8412105-132d-4656-8fd8-296cbd8a1ace" (UID: "d8412105-132d-4656-8fd8-296cbd8a1ace"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:12:00 crc kubenswrapper[4840]: I0930 14:12:00.651680 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8412105-132d-4656-8fd8-296cbd8a1ace-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:00 crc kubenswrapper[4840]: I0930 14:12:00.651719 4840 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d8412105-132d-4656-8fd8-296cbd8a1ace-swiftconf\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:00 crc kubenswrapper[4840]: I0930 14:12:00.651731 4840 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d8412105-132d-4656-8fd8-296cbd8a1ace-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:00 crc kubenswrapper[4840]: I0930 14:12:00.651743 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6cpbv\" (UniqueName: \"kubernetes.io/projected/d8412105-132d-4656-8fd8-296cbd8a1ace-kube-api-access-6cpbv\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:00 crc kubenswrapper[4840]: I0930 14:12:00.651757 4840 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d8412105-132d-4656-8fd8-296cbd8a1ace-dispersionconf\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:00 crc kubenswrapper[4840]: I0930 14:12:00.651769 4840 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d8412105-132d-4656-8fd8-296cbd8a1ace-ring-data-devices\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:00 crc kubenswrapper[4840]: I0930 14:12:00.651781 4840 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d8412105-132d-4656-8fd8-296cbd8a1ace-etc-swift\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:01 crc kubenswrapper[4840]: I0930 14:12:01.018775 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-txwn9" event={"ID":"d8412105-132d-4656-8fd8-296cbd8a1ace","Type":"ContainerDied","Data":"506501210f21c43d3c0d61e0e9471df91db2581c5dcd124819fb55fbbb361891"} Sep 30 14:12:01 crc kubenswrapper[4840]: I0930 14:12:01.018824 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="506501210f21c43d3c0d61e0e9471df91db2581c5dcd124819fb55fbbb361891" Sep 30 14:12:01 crc kubenswrapper[4840]: I0930 14:12:01.019298 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-txwn9" Sep 30 14:12:01 crc kubenswrapper[4840]: I0930 14:12:01.020318 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ae90-account-create-4schj" event={"ID":"b5f49283-d818-4180-9b52-4f05352586e4","Type":"ContainerStarted","Data":"7174d3f5359c27bba905c8bd438c5b90b4e3c82c462022c4de79109a92637736"} Sep 30 14:12:02 crc kubenswrapper[4840]: I0930 14:12:02.036153 4840 generic.go:334] "Generic (PLEG): container finished" podID="b8c84ed0-e20f-445e-9548-4ce979c23476" containerID="d2cdd432cff6329b6a00d90cea6d034e25aced440092be6ed044b88f6112f277" exitCode=0 Sep 30 14:12:02 crc kubenswrapper[4840]: I0930 14:12:02.036216 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b8c84ed0-e20f-445e-9548-4ce979c23476","Type":"ContainerDied","Data":"d2cdd432cff6329b6a00d90cea6d034e25aced440092be6ed044b88f6112f277"} Sep 30 14:12:02 crc kubenswrapper[4840]: I0930 14:12:02.039279 4840 generic.go:334] "Generic (PLEG): container finished" podID="c15070b3-f247-4879-a7dc-618faf7e6e35" containerID="10b078a915c98d66021e33eaf12b5a5033f297335c7c69779a5f20a19252eb1c" exitCode=0 Sep 30 14:12:02 crc kubenswrapper[4840]: I0930 14:12:02.039327 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c15070b3-f247-4879-a7dc-618faf7e6e35","Type":"ContainerDied","Data":"10b078a915c98d66021e33eaf12b5a5033f297335c7c69779a5f20a19252eb1c"} Sep 30 14:12:03 crc kubenswrapper[4840]: I0930 14:12:03.049105 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c15070b3-f247-4879-a7dc-618faf7e6e35","Type":"ContainerStarted","Data":"d4a8b19c278d5dd784385c0c8a86fb6c872488a691d678942412c006d4dd7154"} Sep 30 14:12:03 crc kubenswrapper[4840]: I0930 14:12:03.050220 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:12:03 crc kubenswrapper[4840]: I0930 14:12:03.052409 4840 generic.go:334] "Generic (PLEG): container finished" podID="b5f49283-d818-4180-9b52-4f05352586e4" containerID="493c705051312d1d93803dae118e3980ab83642f89ca3f4ed58c13e68c039320" exitCode=0 Sep 30 14:12:03 crc kubenswrapper[4840]: I0930 14:12:03.052477 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ae90-account-create-4schj" event={"ID":"b5f49283-d818-4180-9b52-4f05352586e4","Type":"ContainerDied","Data":"493c705051312d1d93803dae118e3980ab83642f89ca3f4ed58c13e68c039320"} Sep 30 14:12:03 crc kubenswrapper[4840]: I0930 14:12:03.054654 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b8c84ed0-e20f-445e-9548-4ce979c23476","Type":"ContainerStarted","Data":"ae16b156af4a276de23012df6e733c3b5d485fee9106ae2949eef5807da82adb"} Sep 30 14:12:03 crc kubenswrapper[4840]: I0930 14:12:03.054841 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Sep 30 14:12:03 crc kubenswrapper[4840]: I0930 14:12:03.075925 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=44.234974922 podStartE2EDuration="54.075906681s" podCreationTimestamp="2025-09-30 14:11:09 +0000 UTC" firstStartedPulling="2025-09-30 14:11:20.90302405 +0000 UTC m=+909.532110473" lastFinishedPulling="2025-09-30 14:11:30.743955809 +0000 UTC m=+919.373042232" observedRunningTime="2025-09-30 14:12:03.071690771 +0000 UTC m=+951.700777274" watchObservedRunningTime="2025-09-30 14:12:03.075906681 +0000 UTC m=+951.704993114" Sep 30 14:12:03 crc kubenswrapper[4840]: I0930 14:12:03.111585 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=44.916281137 podStartE2EDuration="54.111566038s" podCreationTimestamp="2025-09-30 14:11:09 +0000 UTC" firstStartedPulling="2025-09-30 14:11:20.895566246 +0000 UTC m=+909.524652659" lastFinishedPulling="2025-09-30 14:11:30.090851137 +0000 UTC m=+918.719937560" observedRunningTime="2025-09-30 14:12:03.106587908 +0000 UTC m=+951.735674361" watchObservedRunningTime="2025-09-30 14:12:03.111566038 +0000 UTC m=+951.740652461" Sep 30 14:12:03 crc kubenswrapper[4840]: I0930 14:12:03.196432 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/66a7a4e6-dd49-4589-bd88-2f5fdf0af70f-etc-swift\") pod \"swift-storage-0\" (UID: \"66a7a4e6-dd49-4589-bd88-2f5fdf0af70f\") " pod="openstack/swift-storage-0" Sep 30 14:12:03 crc kubenswrapper[4840]: I0930 14:12:03.202871 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/66a7a4e6-dd49-4589-bd88-2f5fdf0af70f-etc-swift\") pod \"swift-storage-0\" (UID: \"66a7a4e6-dd49-4589-bd88-2f5fdf0af70f\") " pod="openstack/swift-storage-0" Sep 30 14:12:03 crc kubenswrapper[4840]: I0930 14:12:03.261838 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Sep 30 14:12:03 crc kubenswrapper[4840]: I0930 14:12:03.811386 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Sep 30 14:12:03 crc kubenswrapper[4840]: W0930 14:12:03.826813 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66a7a4e6_dd49_4589_bd88_2f5fdf0af70f.slice/crio-fef31ede1182d89c889588b7555bbf1cc833be4ad3df9dcaf819bd76c8b245e7 WatchSource:0}: Error finding container fef31ede1182d89c889588b7555bbf1cc833be4ad3df9dcaf819bd76c8b245e7: Status 404 returned error can't find the container with id fef31ede1182d89c889588b7555bbf1cc833be4ad3df9dcaf819bd76c8b245e7 Sep 30 14:12:03 crc kubenswrapper[4840]: I0930 14:12:03.902219 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-1a60-account-create-wnsbv"] Sep 30 14:12:03 crc kubenswrapper[4840]: E0930 14:12:03.902583 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8412105-132d-4656-8fd8-296cbd8a1ace" containerName="swift-ring-rebalance" Sep 30 14:12:03 crc kubenswrapper[4840]: I0930 14:12:03.902601 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8412105-132d-4656-8fd8-296cbd8a1ace" containerName="swift-ring-rebalance" Sep 30 14:12:03 crc kubenswrapper[4840]: I0930 14:12:03.902759 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8412105-132d-4656-8fd8-296cbd8a1ace" containerName="swift-ring-rebalance" Sep 30 14:12:03 crc kubenswrapper[4840]: I0930 14:12:03.903698 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-1a60-account-create-wnsbv" Sep 30 14:12:03 crc kubenswrapper[4840]: I0930 14:12:03.907504 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Sep 30 14:12:03 crc kubenswrapper[4840]: I0930 14:12:03.911161 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-1a60-account-create-wnsbv"] Sep 30 14:12:04 crc kubenswrapper[4840]: I0930 14:12:04.011233 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sc992\" (UniqueName: \"kubernetes.io/projected/36cd5d4f-6ef2-440f-a9b3-56610a475114-kube-api-access-sc992\") pod \"keystone-1a60-account-create-wnsbv\" (UID: \"36cd5d4f-6ef2-440f-a9b3-56610a475114\") " pod="openstack/keystone-1a60-account-create-wnsbv" Sep 30 14:12:04 crc kubenswrapper[4840]: I0930 14:12:04.062412 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"66a7a4e6-dd49-4589-bd88-2f5fdf0af70f","Type":"ContainerStarted","Data":"fef31ede1182d89c889588b7555bbf1cc833be4ad3df9dcaf819bd76c8b245e7"} Sep 30 14:12:04 crc kubenswrapper[4840]: I0930 14:12:04.092661 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-4e44-account-create-dcfn6"] Sep 30 14:12:04 crc kubenswrapper[4840]: I0930 14:12:04.093572 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-4e44-account-create-dcfn6" Sep 30 14:12:04 crc kubenswrapper[4840]: I0930 14:12:04.096137 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Sep 30 14:12:04 crc kubenswrapper[4840]: I0930 14:12:04.103868 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-4e44-account-create-dcfn6"] Sep 30 14:12:04 crc kubenswrapper[4840]: I0930 14:12:04.121475 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sc992\" (UniqueName: \"kubernetes.io/projected/36cd5d4f-6ef2-440f-a9b3-56610a475114-kube-api-access-sc992\") pod \"keystone-1a60-account-create-wnsbv\" (UID: \"36cd5d4f-6ef2-440f-a9b3-56610a475114\") " pod="openstack/keystone-1a60-account-create-wnsbv" Sep 30 14:12:04 crc kubenswrapper[4840]: I0930 14:12:04.149292 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sc992\" (UniqueName: \"kubernetes.io/projected/36cd5d4f-6ef2-440f-a9b3-56610a475114-kube-api-access-sc992\") pod \"keystone-1a60-account-create-wnsbv\" (UID: \"36cd5d4f-6ef2-440f-a9b3-56610a475114\") " pod="openstack/keystone-1a60-account-create-wnsbv" Sep 30 14:12:04 crc kubenswrapper[4840]: I0930 14:12:04.222866 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psd5l\" (UniqueName: \"kubernetes.io/projected/fc1ccb29-1c0d-40a2-ae2e-c277fff08748-kube-api-access-psd5l\") pod \"placement-4e44-account-create-dcfn6\" (UID: \"fc1ccb29-1c0d-40a2-ae2e-c277fff08748\") " pod="openstack/placement-4e44-account-create-dcfn6" Sep 30 14:12:04 crc kubenswrapper[4840]: I0930 14:12:04.224186 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-1a60-account-create-wnsbv" Sep 30 14:12:04 crc kubenswrapper[4840]: I0930 14:12:04.324213 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psd5l\" (UniqueName: \"kubernetes.io/projected/fc1ccb29-1c0d-40a2-ae2e-c277fff08748-kube-api-access-psd5l\") pod \"placement-4e44-account-create-dcfn6\" (UID: \"fc1ccb29-1c0d-40a2-ae2e-c277fff08748\") " pod="openstack/placement-4e44-account-create-dcfn6" Sep 30 14:12:04 crc kubenswrapper[4840]: I0930 14:12:04.366124 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psd5l\" (UniqueName: \"kubernetes.io/projected/fc1ccb29-1c0d-40a2-ae2e-c277fff08748-kube-api-access-psd5l\") pod \"placement-4e44-account-create-dcfn6\" (UID: \"fc1ccb29-1c0d-40a2-ae2e-c277fff08748\") " pod="openstack/placement-4e44-account-create-dcfn6" Sep 30 14:12:04 crc kubenswrapper[4840]: I0930 14:12:04.422663 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-4e44-account-create-dcfn6" Sep 30 14:12:04 crc kubenswrapper[4840]: I0930 14:12:04.433474 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ae90-account-create-4schj" Sep 30 14:12:04 crc kubenswrapper[4840]: I0930 14:12:04.526348 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fn7lz\" (UniqueName: \"kubernetes.io/projected/b5f49283-d818-4180-9b52-4f05352586e4-kube-api-access-fn7lz\") pod \"b5f49283-d818-4180-9b52-4f05352586e4\" (UID: \"b5f49283-d818-4180-9b52-4f05352586e4\") " Sep 30 14:12:04 crc kubenswrapper[4840]: I0930 14:12:04.531432 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5f49283-d818-4180-9b52-4f05352586e4-kube-api-access-fn7lz" (OuterVolumeSpecName: "kube-api-access-fn7lz") pod "b5f49283-d818-4180-9b52-4f05352586e4" (UID: "b5f49283-d818-4180-9b52-4f05352586e4"). InnerVolumeSpecName "kube-api-access-fn7lz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:12:04 crc kubenswrapper[4840]: I0930 14:12:04.628218 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fn7lz\" (UniqueName: \"kubernetes.io/projected/b5f49283-d818-4180-9b52-4f05352586e4-kube-api-access-fn7lz\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:04 crc kubenswrapper[4840]: I0930 14:12:04.722291 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-1a60-account-create-wnsbv"] Sep 30 14:12:04 crc kubenswrapper[4840]: W0930 14:12:04.834356 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod36cd5d4f_6ef2_440f_a9b3_56610a475114.slice/crio-f73a8b88ab90228ab283490c7187bb1b6789daab2a815c25becbf2537d982ece WatchSource:0}: Error finding container f73a8b88ab90228ab283490c7187bb1b6789daab2a815c25becbf2537d982ece: Status 404 returned error can't find the container with id f73a8b88ab90228ab283490c7187bb1b6789daab2a815c25becbf2537d982ece Sep 30 14:12:04 crc kubenswrapper[4840]: I0930 14:12:04.862813 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-4e44-account-create-dcfn6"] Sep 30 14:12:04 crc kubenswrapper[4840]: W0930 14:12:04.884020 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfc1ccb29_1c0d_40a2_ae2e_c277fff08748.slice/crio-569e06ceb1c6ea1438e0247a985ebf14bbb244068a371b13e98a779f9ce7e051 WatchSource:0}: Error finding container 569e06ceb1c6ea1438e0247a985ebf14bbb244068a371b13e98a779f9ce7e051: Status 404 returned error can't find the container with id 569e06ceb1c6ea1438e0247a985ebf14bbb244068a371b13e98a779f9ce7e051 Sep 30 14:12:05 crc kubenswrapper[4840]: I0930 14:12:05.079313 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-4e44-account-create-dcfn6" event={"ID":"fc1ccb29-1c0d-40a2-ae2e-c277fff08748","Type":"ContainerStarted","Data":"5cd9591b125dcec7aba5800e5f1bfd0c285ca6e095931a0566197f1f2764c825"} Sep 30 14:12:05 crc kubenswrapper[4840]: I0930 14:12:05.079356 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-4e44-account-create-dcfn6" event={"ID":"fc1ccb29-1c0d-40a2-ae2e-c277fff08748","Type":"ContainerStarted","Data":"569e06ceb1c6ea1438e0247a985ebf14bbb244068a371b13e98a779f9ce7e051"} Sep 30 14:12:05 crc kubenswrapper[4840]: I0930 14:12:05.080909 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"66a7a4e6-dd49-4589-bd88-2f5fdf0af70f","Type":"ContainerStarted","Data":"3abefd648d7d93b1b9d02424a14d85ed8f01683529a839728390111b263d9930"} Sep 30 14:12:05 crc kubenswrapper[4840]: I0930 14:12:05.083517 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ae90-account-create-4schj" event={"ID":"b5f49283-d818-4180-9b52-4f05352586e4","Type":"ContainerDied","Data":"7174d3f5359c27bba905c8bd438c5b90b4e3c82c462022c4de79109a92637736"} Sep 30 14:12:05 crc kubenswrapper[4840]: I0930 14:12:05.083606 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7174d3f5359c27bba905c8bd438c5b90b4e3c82c462022c4de79109a92637736" Sep 30 14:12:05 crc kubenswrapper[4840]: I0930 14:12:05.083525 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ae90-account-create-4schj" Sep 30 14:12:05 crc kubenswrapper[4840]: I0930 14:12:05.085197 4840 generic.go:334] "Generic (PLEG): container finished" podID="36cd5d4f-6ef2-440f-a9b3-56610a475114" containerID="5bac448d393c5d2cdabbf98690a456182f58a8e6161c29658b38fb96e3b3bea1" exitCode=0 Sep 30 14:12:05 crc kubenswrapper[4840]: I0930 14:12:05.085224 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-1a60-account-create-wnsbv" event={"ID":"36cd5d4f-6ef2-440f-a9b3-56610a475114","Type":"ContainerDied","Data":"5bac448d393c5d2cdabbf98690a456182f58a8e6161c29658b38fb96e3b3bea1"} Sep 30 14:12:05 crc kubenswrapper[4840]: I0930 14:12:05.085239 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-1a60-account-create-wnsbv" event={"ID":"36cd5d4f-6ef2-440f-a9b3-56610a475114","Type":"ContainerStarted","Data":"f73a8b88ab90228ab283490c7187bb1b6789daab2a815c25becbf2537d982ece"} Sep 30 14:12:05 crc kubenswrapper[4840]: I0930 14:12:05.109160 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-4e44-account-create-dcfn6" podStartSLOduration=1.109141587 podStartE2EDuration="1.109141587s" podCreationTimestamp="2025-09-30 14:12:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:12:05.095827971 +0000 UTC m=+953.724914404" watchObservedRunningTime="2025-09-30 14:12:05.109141587 +0000 UTC m=+953.738228020" Sep 30 14:12:05 crc kubenswrapper[4840]: I0930 14:12:05.278191 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-fc5jg" Sep 30 14:12:05 crc kubenswrapper[4840]: I0930 14:12:05.279885 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-cngvz" podUID="e8a63b50-ecd5-4993-a890-8c94bc2d5e60" containerName="ovn-controller" probeResult="failure" output=< Sep 30 14:12:05 crc kubenswrapper[4840]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Sep 30 14:12:05 crc kubenswrapper[4840]: > Sep 30 14:12:05 crc kubenswrapper[4840]: I0930 14:12:05.289133 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-fc5jg" Sep 30 14:12:05 crc kubenswrapper[4840]: I0930 14:12:05.506844 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-cngvz-config-lbvv7"] Sep 30 14:12:05 crc kubenswrapper[4840]: E0930 14:12:05.507462 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5f49283-d818-4180-9b52-4f05352586e4" containerName="mariadb-account-create" Sep 30 14:12:05 crc kubenswrapper[4840]: I0930 14:12:05.507482 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5f49283-d818-4180-9b52-4f05352586e4" containerName="mariadb-account-create" Sep 30 14:12:05 crc kubenswrapper[4840]: I0930 14:12:05.507686 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5f49283-d818-4180-9b52-4f05352586e4" containerName="mariadb-account-create" Sep 30 14:12:05 crc kubenswrapper[4840]: I0930 14:12:05.508176 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-cngvz-config-lbvv7" Sep 30 14:12:05 crc kubenswrapper[4840]: I0930 14:12:05.512376 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Sep 30 14:12:05 crc kubenswrapper[4840]: I0930 14:12:05.525289 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-cngvz-config-lbvv7"] Sep 30 14:12:05 crc kubenswrapper[4840]: I0930 14:12:05.646674 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17170628-abe7-419e-9d8e-0340faa97e88-scripts\") pod \"ovn-controller-cngvz-config-lbvv7\" (UID: \"17170628-abe7-419e-9d8e-0340faa97e88\") " pod="openstack/ovn-controller-cngvz-config-lbvv7" Sep 30 14:12:05 crc kubenswrapper[4840]: I0930 14:12:05.646914 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/17170628-abe7-419e-9d8e-0340faa97e88-var-run-ovn\") pod \"ovn-controller-cngvz-config-lbvv7\" (UID: \"17170628-abe7-419e-9d8e-0340faa97e88\") " pod="openstack/ovn-controller-cngvz-config-lbvv7" Sep 30 14:12:05 crc kubenswrapper[4840]: I0930 14:12:05.647122 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/17170628-abe7-419e-9d8e-0340faa97e88-var-log-ovn\") pod \"ovn-controller-cngvz-config-lbvv7\" (UID: \"17170628-abe7-419e-9d8e-0340faa97e88\") " pod="openstack/ovn-controller-cngvz-config-lbvv7" Sep 30 14:12:05 crc kubenswrapper[4840]: I0930 14:12:05.647307 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/17170628-abe7-419e-9d8e-0340faa97e88-additional-scripts\") pod \"ovn-controller-cngvz-config-lbvv7\" (UID: \"17170628-abe7-419e-9d8e-0340faa97e88\") " pod="openstack/ovn-controller-cngvz-config-lbvv7" Sep 30 14:12:05 crc kubenswrapper[4840]: I0930 14:12:05.647478 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/17170628-abe7-419e-9d8e-0340faa97e88-var-run\") pod \"ovn-controller-cngvz-config-lbvv7\" (UID: \"17170628-abe7-419e-9d8e-0340faa97e88\") " pod="openstack/ovn-controller-cngvz-config-lbvv7" Sep 30 14:12:05 crc kubenswrapper[4840]: I0930 14:12:05.647542 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6cf5\" (UniqueName: \"kubernetes.io/projected/17170628-abe7-419e-9d8e-0340faa97e88-kube-api-access-v6cf5\") pod \"ovn-controller-cngvz-config-lbvv7\" (UID: \"17170628-abe7-419e-9d8e-0340faa97e88\") " pod="openstack/ovn-controller-cngvz-config-lbvv7" Sep 30 14:12:05 crc kubenswrapper[4840]: I0930 14:12:05.749325 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/17170628-abe7-419e-9d8e-0340faa97e88-var-log-ovn\") pod \"ovn-controller-cngvz-config-lbvv7\" (UID: \"17170628-abe7-419e-9d8e-0340faa97e88\") " pod="openstack/ovn-controller-cngvz-config-lbvv7" Sep 30 14:12:05 crc kubenswrapper[4840]: I0930 14:12:05.749410 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/17170628-abe7-419e-9d8e-0340faa97e88-additional-scripts\") pod \"ovn-controller-cngvz-config-lbvv7\" (UID: \"17170628-abe7-419e-9d8e-0340faa97e88\") " pod="openstack/ovn-controller-cngvz-config-lbvv7" Sep 30 14:12:05 crc kubenswrapper[4840]: I0930 14:12:05.749469 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/17170628-abe7-419e-9d8e-0340faa97e88-var-run\") pod \"ovn-controller-cngvz-config-lbvv7\" (UID: \"17170628-abe7-419e-9d8e-0340faa97e88\") " pod="openstack/ovn-controller-cngvz-config-lbvv7" Sep 30 14:12:05 crc kubenswrapper[4840]: I0930 14:12:05.749494 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6cf5\" (UniqueName: \"kubernetes.io/projected/17170628-abe7-419e-9d8e-0340faa97e88-kube-api-access-v6cf5\") pod \"ovn-controller-cngvz-config-lbvv7\" (UID: \"17170628-abe7-419e-9d8e-0340faa97e88\") " pod="openstack/ovn-controller-cngvz-config-lbvv7" Sep 30 14:12:05 crc kubenswrapper[4840]: I0930 14:12:05.749541 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17170628-abe7-419e-9d8e-0340faa97e88-scripts\") pod \"ovn-controller-cngvz-config-lbvv7\" (UID: \"17170628-abe7-419e-9d8e-0340faa97e88\") " pod="openstack/ovn-controller-cngvz-config-lbvv7" Sep 30 14:12:05 crc kubenswrapper[4840]: I0930 14:12:05.749599 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/17170628-abe7-419e-9d8e-0340faa97e88-var-run-ovn\") pod \"ovn-controller-cngvz-config-lbvv7\" (UID: \"17170628-abe7-419e-9d8e-0340faa97e88\") " pod="openstack/ovn-controller-cngvz-config-lbvv7" Sep 30 14:12:05 crc kubenswrapper[4840]: I0930 14:12:05.749664 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/17170628-abe7-419e-9d8e-0340faa97e88-var-log-ovn\") pod \"ovn-controller-cngvz-config-lbvv7\" (UID: \"17170628-abe7-419e-9d8e-0340faa97e88\") " pod="openstack/ovn-controller-cngvz-config-lbvv7" Sep 30 14:12:05 crc kubenswrapper[4840]: I0930 14:12:05.749705 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/17170628-abe7-419e-9d8e-0340faa97e88-var-run-ovn\") pod \"ovn-controller-cngvz-config-lbvv7\" (UID: \"17170628-abe7-419e-9d8e-0340faa97e88\") " pod="openstack/ovn-controller-cngvz-config-lbvv7" Sep 30 14:12:05 crc kubenswrapper[4840]: I0930 14:12:05.749708 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/17170628-abe7-419e-9d8e-0340faa97e88-var-run\") pod \"ovn-controller-cngvz-config-lbvv7\" (UID: \"17170628-abe7-419e-9d8e-0340faa97e88\") " pod="openstack/ovn-controller-cngvz-config-lbvv7" Sep 30 14:12:05 crc kubenswrapper[4840]: I0930 14:12:05.750302 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/17170628-abe7-419e-9d8e-0340faa97e88-additional-scripts\") pod \"ovn-controller-cngvz-config-lbvv7\" (UID: \"17170628-abe7-419e-9d8e-0340faa97e88\") " pod="openstack/ovn-controller-cngvz-config-lbvv7" Sep 30 14:12:05 crc kubenswrapper[4840]: I0930 14:12:05.752010 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17170628-abe7-419e-9d8e-0340faa97e88-scripts\") pod \"ovn-controller-cngvz-config-lbvv7\" (UID: \"17170628-abe7-419e-9d8e-0340faa97e88\") " pod="openstack/ovn-controller-cngvz-config-lbvv7" Sep 30 14:12:05 crc kubenswrapper[4840]: I0930 14:12:05.771181 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6cf5\" (UniqueName: \"kubernetes.io/projected/17170628-abe7-419e-9d8e-0340faa97e88-kube-api-access-v6cf5\") pod \"ovn-controller-cngvz-config-lbvv7\" (UID: \"17170628-abe7-419e-9d8e-0340faa97e88\") " pod="openstack/ovn-controller-cngvz-config-lbvv7" Sep 30 14:12:05 crc kubenswrapper[4840]: I0930 14:12:05.849367 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-cngvz-config-lbvv7" Sep 30 14:12:06 crc kubenswrapper[4840]: I0930 14:12:06.095909 4840 generic.go:334] "Generic (PLEG): container finished" podID="fc1ccb29-1c0d-40a2-ae2e-c277fff08748" containerID="5cd9591b125dcec7aba5800e5f1bfd0c285ca6e095931a0566197f1f2764c825" exitCode=0 Sep 30 14:12:06 crc kubenswrapper[4840]: I0930 14:12:06.096097 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-4e44-account-create-dcfn6" event={"ID":"fc1ccb29-1c0d-40a2-ae2e-c277fff08748","Type":"ContainerDied","Data":"5cd9591b125dcec7aba5800e5f1bfd0c285ca6e095931a0566197f1f2764c825"} Sep 30 14:12:06 crc kubenswrapper[4840]: I0930 14:12:06.100211 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"66a7a4e6-dd49-4589-bd88-2f5fdf0af70f","Type":"ContainerStarted","Data":"5d748c410f5c58fdece87a6b98141d6c681f5364ea9b50750f9aabf8373fcba6"} Sep 30 14:12:06 crc kubenswrapper[4840]: I0930 14:12:06.100252 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"66a7a4e6-dd49-4589-bd88-2f5fdf0af70f","Type":"ContainerStarted","Data":"8916ff28deab579c3324c80c367325330480624715062ffbef8b6b16a57ff75e"} Sep 30 14:12:06 crc kubenswrapper[4840]: I0930 14:12:06.100266 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"66a7a4e6-dd49-4589-bd88-2f5fdf0af70f","Type":"ContainerStarted","Data":"223d0f0ec43b70846c42c87e2f1a926f031c92ee9f6fad4bc479c865a56eb853"} Sep 30 14:12:06 crc kubenswrapper[4840]: I0930 14:12:06.329458 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-cngvz-config-lbvv7"] Sep 30 14:12:06 crc kubenswrapper[4840]: W0930 14:12:06.337599 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17170628_abe7_419e_9d8e_0340faa97e88.slice/crio-4082ccfaeb47651a9a7ed290e02499b76024e6f7ef15eea78115d85c99edd89c WatchSource:0}: Error finding container 4082ccfaeb47651a9a7ed290e02499b76024e6f7ef15eea78115d85c99edd89c: Status 404 returned error can't find the container with id 4082ccfaeb47651a9a7ed290e02499b76024e6f7ef15eea78115d85c99edd89c Sep 30 14:12:06 crc kubenswrapper[4840]: I0930 14:12:06.450646 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-1a60-account-create-wnsbv" Sep 30 14:12:06 crc kubenswrapper[4840]: I0930 14:12:06.561250 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sc992\" (UniqueName: \"kubernetes.io/projected/36cd5d4f-6ef2-440f-a9b3-56610a475114-kube-api-access-sc992\") pod \"36cd5d4f-6ef2-440f-a9b3-56610a475114\" (UID: \"36cd5d4f-6ef2-440f-a9b3-56610a475114\") " Sep 30 14:12:06 crc kubenswrapper[4840]: I0930 14:12:06.567443 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36cd5d4f-6ef2-440f-a9b3-56610a475114-kube-api-access-sc992" (OuterVolumeSpecName: "kube-api-access-sc992") pod "36cd5d4f-6ef2-440f-a9b3-56610a475114" (UID: "36cd5d4f-6ef2-440f-a9b3-56610a475114"). InnerVolumeSpecName "kube-api-access-sc992". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:12:06 crc kubenswrapper[4840]: I0930 14:12:06.663641 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sc992\" (UniqueName: \"kubernetes.io/projected/36cd5d4f-6ef2-440f-a9b3-56610a475114-kube-api-access-sc992\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:07 crc kubenswrapper[4840]: I0930 14:12:07.108535 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-1a60-account-create-wnsbv" Sep 30 14:12:07 crc kubenswrapper[4840]: I0930 14:12:07.108533 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-1a60-account-create-wnsbv" event={"ID":"36cd5d4f-6ef2-440f-a9b3-56610a475114","Type":"ContainerDied","Data":"f73a8b88ab90228ab283490c7187bb1b6789daab2a815c25becbf2537d982ece"} Sep 30 14:12:07 crc kubenswrapper[4840]: I0930 14:12:07.109004 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f73a8b88ab90228ab283490c7187bb1b6789daab2a815c25becbf2537d982ece" Sep 30 14:12:07 crc kubenswrapper[4840]: I0930 14:12:07.112302 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-cngvz-config-lbvv7" event={"ID":"17170628-abe7-419e-9d8e-0340faa97e88","Type":"ContainerStarted","Data":"9bbf388fd07ba823ead856edc82d164211722304b951e16bdab0ea953989ade1"} Sep 30 14:12:07 crc kubenswrapper[4840]: I0930 14:12:07.112358 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-cngvz-config-lbvv7" event={"ID":"17170628-abe7-419e-9d8e-0340faa97e88","Type":"ContainerStarted","Data":"4082ccfaeb47651a9a7ed290e02499b76024e6f7ef15eea78115d85c99edd89c"} Sep 30 14:12:07 crc kubenswrapper[4840]: I0930 14:12:07.422355 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-4e44-account-create-dcfn6" Sep 30 14:12:07 crc kubenswrapper[4840]: I0930 14:12:07.578482 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-psd5l\" (UniqueName: \"kubernetes.io/projected/fc1ccb29-1c0d-40a2-ae2e-c277fff08748-kube-api-access-psd5l\") pod \"fc1ccb29-1c0d-40a2-ae2e-c277fff08748\" (UID: \"fc1ccb29-1c0d-40a2-ae2e-c277fff08748\") " Sep 30 14:12:07 crc kubenswrapper[4840]: I0930 14:12:07.586478 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc1ccb29-1c0d-40a2-ae2e-c277fff08748-kube-api-access-psd5l" (OuterVolumeSpecName: "kube-api-access-psd5l") pod "fc1ccb29-1c0d-40a2-ae2e-c277fff08748" (UID: "fc1ccb29-1c0d-40a2-ae2e-c277fff08748"). InnerVolumeSpecName "kube-api-access-psd5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:12:07 crc kubenswrapper[4840]: I0930 14:12:07.680167 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-psd5l\" (UniqueName: \"kubernetes.io/projected/fc1ccb29-1c0d-40a2-ae2e-c277fff08748-kube-api-access-psd5l\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:08 crc kubenswrapper[4840]: I0930 14:12:08.133727 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-4e44-account-create-dcfn6" event={"ID":"fc1ccb29-1c0d-40a2-ae2e-c277fff08748","Type":"ContainerDied","Data":"569e06ceb1c6ea1438e0247a985ebf14bbb244068a371b13e98a779f9ce7e051"} Sep 30 14:12:08 crc kubenswrapper[4840]: I0930 14:12:08.134064 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="569e06ceb1c6ea1438e0247a985ebf14bbb244068a371b13e98a779f9ce7e051" Sep 30 14:12:08 crc kubenswrapper[4840]: I0930 14:12:08.134137 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-4e44-account-create-dcfn6" Sep 30 14:12:08 crc kubenswrapper[4840]: I0930 14:12:08.139679 4840 generic.go:334] "Generic (PLEG): container finished" podID="17170628-abe7-419e-9d8e-0340faa97e88" containerID="9bbf388fd07ba823ead856edc82d164211722304b951e16bdab0ea953989ade1" exitCode=0 Sep 30 14:12:08 crc kubenswrapper[4840]: I0930 14:12:08.139751 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-cngvz-config-lbvv7" event={"ID":"17170628-abe7-419e-9d8e-0340faa97e88","Type":"ContainerDied","Data":"9bbf388fd07ba823ead856edc82d164211722304b951e16bdab0ea953989ade1"} Sep 30 14:12:08 crc kubenswrapper[4840]: I0930 14:12:08.144146 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"66a7a4e6-dd49-4589-bd88-2f5fdf0af70f","Type":"ContainerStarted","Data":"8acc114c6af4bcce99dccae262639672a12221c17ba385b90e4054a15ec93d6a"} Sep 30 14:12:08 crc kubenswrapper[4840]: I0930 14:12:08.450455 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-cngvz-config-lbvv7" Sep 30 14:12:08 crc kubenswrapper[4840]: I0930 14:12:08.601017 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17170628-abe7-419e-9d8e-0340faa97e88-scripts\") pod \"17170628-abe7-419e-9d8e-0340faa97e88\" (UID: \"17170628-abe7-419e-9d8e-0340faa97e88\") " Sep 30 14:12:08 crc kubenswrapper[4840]: I0930 14:12:08.601078 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/17170628-abe7-419e-9d8e-0340faa97e88-var-run-ovn\") pod \"17170628-abe7-419e-9d8e-0340faa97e88\" (UID: \"17170628-abe7-419e-9d8e-0340faa97e88\") " Sep 30 14:12:08 crc kubenswrapper[4840]: I0930 14:12:08.601118 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/17170628-abe7-419e-9d8e-0340faa97e88-var-run\") pod \"17170628-abe7-419e-9d8e-0340faa97e88\" (UID: \"17170628-abe7-419e-9d8e-0340faa97e88\") " Sep 30 14:12:08 crc kubenswrapper[4840]: I0930 14:12:08.601140 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/17170628-abe7-419e-9d8e-0340faa97e88-var-log-ovn\") pod \"17170628-abe7-419e-9d8e-0340faa97e88\" (UID: \"17170628-abe7-419e-9d8e-0340faa97e88\") " Sep 30 14:12:08 crc kubenswrapper[4840]: I0930 14:12:08.601196 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/17170628-abe7-419e-9d8e-0340faa97e88-additional-scripts\") pod \"17170628-abe7-419e-9d8e-0340faa97e88\" (UID: \"17170628-abe7-419e-9d8e-0340faa97e88\") " Sep 30 14:12:08 crc kubenswrapper[4840]: I0930 14:12:08.601217 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/17170628-abe7-419e-9d8e-0340faa97e88-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "17170628-abe7-419e-9d8e-0340faa97e88" (UID: "17170628-abe7-419e-9d8e-0340faa97e88"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:12:08 crc kubenswrapper[4840]: I0930 14:12:08.601244 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/17170628-abe7-419e-9d8e-0340faa97e88-var-run" (OuterVolumeSpecName: "var-run") pod "17170628-abe7-419e-9d8e-0340faa97e88" (UID: "17170628-abe7-419e-9d8e-0340faa97e88"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:12:08 crc kubenswrapper[4840]: I0930 14:12:08.601275 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6cf5\" (UniqueName: \"kubernetes.io/projected/17170628-abe7-419e-9d8e-0340faa97e88-kube-api-access-v6cf5\") pod \"17170628-abe7-419e-9d8e-0340faa97e88\" (UID: \"17170628-abe7-419e-9d8e-0340faa97e88\") " Sep 30 14:12:08 crc kubenswrapper[4840]: I0930 14:12:08.601272 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/17170628-abe7-419e-9d8e-0340faa97e88-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "17170628-abe7-419e-9d8e-0340faa97e88" (UID: "17170628-abe7-419e-9d8e-0340faa97e88"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:12:08 crc kubenswrapper[4840]: I0930 14:12:08.601564 4840 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/17170628-abe7-419e-9d8e-0340faa97e88-var-run-ovn\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:08 crc kubenswrapper[4840]: I0930 14:12:08.601577 4840 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/17170628-abe7-419e-9d8e-0340faa97e88-var-run\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:08 crc kubenswrapper[4840]: I0930 14:12:08.601586 4840 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/17170628-abe7-419e-9d8e-0340faa97e88-var-log-ovn\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:08 crc kubenswrapper[4840]: I0930 14:12:08.601886 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17170628-abe7-419e-9d8e-0340faa97e88-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "17170628-abe7-419e-9d8e-0340faa97e88" (UID: "17170628-abe7-419e-9d8e-0340faa97e88"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:12:08 crc kubenswrapper[4840]: I0930 14:12:08.602178 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17170628-abe7-419e-9d8e-0340faa97e88-scripts" (OuterVolumeSpecName: "scripts") pod "17170628-abe7-419e-9d8e-0340faa97e88" (UID: "17170628-abe7-419e-9d8e-0340faa97e88"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:12:08 crc kubenswrapper[4840]: I0930 14:12:08.607365 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17170628-abe7-419e-9d8e-0340faa97e88-kube-api-access-v6cf5" (OuterVolumeSpecName: "kube-api-access-v6cf5") pod "17170628-abe7-419e-9d8e-0340faa97e88" (UID: "17170628-abe7-419e-9d8e-0340faa97e88"). InnerVolumeSpecName "kube-api-access-v6cf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:12:08 crc kubenswrapper[4840]: I0930 14:12:08.703678 4840 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/17170628-abe7-419e-9d8e-0340faa97e88-additional-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:08 crc kubenswrapper[4840]: I0930 14:12:08.703725 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6cf5\" (UniqueName: \"kubernetes.io/projected/17170628-abe7-419e-9d8e-0340faa97e88-kube-api-access-v6cf5\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:08 crc kubenswrapper[4840]: I0930 14:12:08.703738 4840 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17170628-abe7-419e-9d8e-0340faa97e88-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:09 crc kubenswrapper[4840]: I0930 14:12:09.153278 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-cngvz-config-lbvv7" event={"ID":"17170628-abe7-419e-9d8e-0340faa97e88","Type":"ContainerDied","Data":"4082ccfaeb47651a9a7ed290e02499b76024e6f7ef15eea78115d85c99edd89c"} Sep 30 14:12:09 crc kubenswrapper[4840]: I0930 14:12:09.153663 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4082ccfaeb47651a9a7ed290e02499b76024e6f7ef15eea78115d85c99edd89c" Sep 30 14:12:09 crc kubenswrapper[4840]: I0930 14:12:09.153337 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-cngvz-config-lbvv7" Sep 30 14:12:09 crc kubenswrapper[4840]: I0930 14:12:09.157625 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"66a7a4e6-dd49-4589-bd88-2f5fdf0af70f","Type":"ContainerStarted","Data":"34b91dd1b5446c8900280f25c9a1bc4c4436dc7ac4c77518ea58eef1642577b7"} Sep 30 14:12:09 crc kubenswrapper[4840]: I0930 14:12:09.157664 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"66a7a4e6-dd49-4589-bd88-2f5fdf0af70f","Type":"ContainerStarted","Data":"bff33e82a95c8206232111318636b35f5b6f69abd4c27ceddd6cad28b89eff08"} Sep 30 14:12:09 crc kubenswrapper[4840]: I0930 14:12:09.157675 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"66a7a4e6-dd49-4589-bd88-2f5fdf0af70f","Type":"ContainerStarted","Data":"5fe12707e28e27be1bc886cb8e3db1d96b37edae07ab2b01c5085f6d7fab0e56"} Sep 30 14:12:09 crc kubenswrapper[4840]: I0930 14:12:09.554644 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-cngvz-config-lbvv7"] Sep 30 14:12:09 crc kubenswrapper[4840]: I0930 14:12:09.561885 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-cngvz-config-lbvv7"] Sep 30 14:12:09 crc kubenswrapper[4840]: I0930 14:12:09.754976 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-czdht"] Sep 30 14:12:09 crc kubenswrapper[4840]: E0930 14:12:09.755305 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17170628-abe7-419e-9d8e-0340faa97e88" containerName="ovn-config" Sep 30 14:12:09 crc kubenswrapper[4840]: I0930 14:12:09.755318 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="17170628-abe7-419e-9d8e-0340faa97e88" containerName="ovn-config" Sep 30 14:12:09 crc kubenswrapper[4840]: E0930 14:12:09.755336 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36cd5d4f-6ef2-440f-a9b3-56610a475114" containerName="mariadb-account-create" Sep 30 14:12:09 crc kubenswrapper[4840]: I0930 14:12:09.755342 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="36cd5d4f-6ef2-440f-a9b3-56610a475114" containerName="mariadb-account-create" Sep 30 14:12:09 crc kubenswrapper[4840]: E0930 14:12:09.755352 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc1ccb29-1c0d-40a2-ae2e-c277fff08748" containerName="mariadb-account-create" Sep 30 14:12:09 crc kubenswrapper[4840]: I0930 14:12:09.755358 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc1ccb29-1c0d-40a2-ae2e-c277fff08748" containerName="mariadb-account-create" Sep 30 14:12:09 crc kubenswrapper[4840]: I0930 14:12:09.755497 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="36cd5d4f-6ef2-440f-a9b3-56610a475114" containerName="mariadb-account-create" Sep 30 14:12:09 crc kubenswrapper[4840]: I0930 14:12:09.755513 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="17170628-abe7-419e-9d8e-0340faa97e88" containerName="ovn-config" Sep 30 14:12:09 crc kubenswrapper[4840]: I0930 14:12:09.755533 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc1ccb29-1c0d-40a2-ae2e-c277fff08748" containerName="mariadb-account-create" Sep 30 14:12:09 crc kubenswrapper[4840]: I0930 14:12:09.756047 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-czdht" Sep 30 14:12:09 crc kubenswrapper[4840]: I0930 14:12:09.759794 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Sep 30 14:12:09 crc kubenswrapper[4840]: I0930 14:12:09.759838 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-7s8kx" Sep 30 14:12:09 crc kubenswrapper[4840]: I0930 14:12:09.776287 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-czdht"] Sep 30 14:12:09 crc kubenswrapper[4840]: I0930 14:12:09.920465 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkk9r\" (UniqueName: \"kubernetes.io/projected/a1f15dc9-8926-4c0c-89da-a27216c4cb1a-kube-api-access-xkk9r\") pod \"glance-db-sync-czdht\" (UID: \"a1f15dc9-8926-4c0c-89da-a27216c4cb1a\") " pod="openstack/glance-db-sync-czdht" Sep 30 14:12:09 crc kubenswrapper[4840]: I0930 14:12:09.920835 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1f15dc9-8926-4c0c-89da-a27216c4cb1a-combined-ca-bundle\") pod \"glance-db-sync-czdht\" (UID: \"a1f15dc9-8926-4c0c-89da-a27216c4cb1a\") " pod="openstack/glance-db-sync-czdht" Sep 30 14:12:09 crc kubenswrapper[4840]: I0930 14:12:09.920934 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a1f15dc9-8926-4c0c-89da-a27216c4cb1a-db-sync-config-data\") pod \"glance-db-sync-czdht\" (UID: \"a1f15dc9-8926-4c0c-89da-a27216c4cb1a\") " pod="openstack/glance-db-sync-czdht" Sep 30 14:12:09 crc kubenswrapper[4840]: I0930 14:12:09.921120 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1f15dc9-8926-4c0c-89da-a27216c4cb1a-config-data\") pod \"glance-db-sync-czdht\" (UID: \"a1f15dc9-8926-4c0c-89da-a27216c4cb1a\") " pod="openstack/glance-db-sync-czdht" Sep 30 14:12:10 crc kubenswrapper[4840]: I0930 14:12:10.023335 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a1f15dc9-8926-4c0c-89da-a27216c4cb1a-db-sync-config-data\") pod \"glance-db-sync-czdht\" (UID: \"a1f15dc9-8926-4c0c-89da-a27216c4cb1a\") " pod="openstack/glance-db-sync-czdht" Sep 30 14:12:10 crc kubenswrapper[4840]: I0930 14:12:10.023435 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1f15dc9-8926-4c0c-89da-a27216c4cb1a-config-data\") pod \"glance-db-sync-czdht\" (UID: \"a1f15dc9-8926-4c0c-89da-a27216c4cb1a\") " pod="openstack/glance-db-sync-czdht" Sep 30 14:12:10 crc kubenswrapper[4840]: I0930 14:12:10.023477 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkk9r\" (UniqueName: \"kubernetes.io/projected/a1f15dc9-8926-4c0c-89da-a27216c4cb1a-kube-api-access-xkk9r\") pod \"glance-db-sync-czdht\" (UID: \"a1f15dc9-8926-4c0c-89da-a27216c4cb1a\") " pod="openstack/glance-db-sync-czdht" Sep 30 14:12:10 crc kubenswrapper[4840]: I0930 14:12:10.023520 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1f15dc9-8926-4c0c-89da-a27216c4cb1a-combined-ca-bundle\") pod \"glance-db-sync-czdht\" (UID: \"a1f15dc9-8926-4c0c-89da-a27216c4cb1a\") " pod="openstack/glance-db-sync-czdht" Sep 30 14:12:10 crc kubenswrapper[4840]: I0930 14:12:10.029928 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1f15dc9-8926-4c0c-89da-a27216c4cb1a-config-data\") pod \"glance-db-sync-czdht\" (UID: \"a1f15dc9-8926-4c0c-89da-a27216c4cb1a\") " pod="openstack/glance-db-sync-czdht" Sep 30 14:12:10 crc kubenswrapper[4840]: I0930 14:12:10.030731 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1f15dc9-8926-4c0c-89da-a27216c4cb1a-combined-ca-bundle\") pod \"glance-db-sync-czdht\" (UID: \"a1f15dc9-8926-4c0c-89da-a27216c4cb1a\") " pod="openstack/glance-db-sync-czdht" Sep 30 14:12:10 crc kubenswrapper[4840]: I0930 14:12:10.031454 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a1f15dc9-8926-4c0c-89da-a27216c4cb1a-db-sync-config-data\") pod \"glance-db-sync-czdht\" (UID: \"a1f15dc9-8926-4c0c-89da-a27216c4cb1a\") " pod="openstack/glance-db-sync-czdht" Sep 30 14:12:10 crc kubenswrapper[4840]: I0930 14:12:10.044466 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkk9r\" (UniqueName: \"kubernetes.io/projected/a1f15dc9-8926-4c0c-89da-a27216c4cb1a-kube-api-access-xkk9r\") pod \"glance-db-sync-czdht\" (UID: \"a1f15dc9-8926-4c0c-89da-a27216c4cb1a\") " pod="openstack/glance-db-sync-czdht" Sep 30 14:12:10 crc kubenswrapper[4840]: I0930 14:12:10.072130 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-czdht" Sep 30 14:12:10 crc kubenswrapper[4840]: I0930 14:12:10.173431 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17170628-abe7-419e-9d8e-0340faa97e88" path="/var/lib/kubelet/pods/17170628-abe7-419e-9d8e-0340faa97e88/volumes" Sep 30 14:12:10 crc kubenswrapper[4840]: I0930 14:12:10.300723 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-cngvz" Sep 30 14:12:10 crc kubenswrapper[4840]: I0930 14:12:10.751107 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-czdht"] Sep 30 14:12:10 crc kubenswrapper[4840]: W0930 14:12:10.760503 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1f15dc9_8926_4c0c_89da_a27216c4cb1a.slice/crio-1b6a70ed22a1c84410ef38d235d326cd24d3a4ec5f6c91f1d4870ebfc878bafc WatchSource:0}: Error finding container 1b6a70ed22a1c84410ef38d235d326cd24d3a4ec5f6c91f1d4870ebfc878bafc: Status 404 returned error can't find the container with id 1b6a70ed22a1c84410ef38d235d326cd24d3a4ec5f6c91f1d4870ebfc878bafc Sep 30 14:12:11 crc kubenswrapper[4840]: I0930 14:12:11.184003 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-czdht" event={"ID":"a1f15dc9-8926-4c0c-89da-a27216c4cb1a","Type":"ContainerStarted","Data":"1b6a70ed22a1c84410ef38d235d326cd24d3a4ec5f6c91f1d4870ebfc878bafc"} Sep 30 14:12:11 crc kubenswrapper[4840]: I0930 14:12:11.191288 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"66a7a4e6-dd49-4589-bd88-2f5fdf0af70f","Type":"ContainerStarted","Data":"445a5c40419ef99df7df3e8fedf473f229583bf17b621232b7d3fc6ee81f7218"} Sep 30 14:12:11 crc kubenswrapper[4840]: I0930 14:12:11.191328 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"66a7a4e6-dd49-4589-bd88-2f5fdf0af70f","Type":"ContainerStarted","Data":"d8aa2c5d22cfbe1e743de611763b3ee88da3794fe6d015357688140ab0f8f3b6"} Sep 30 14:12:11 crc kubenswrapper[4840]: I0930 14:12:11.191339 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"66a7a4e6-dd49-4589-bd88-2f5fdf0af70f","Type":"ContainerStarted","Data":"0f9f69184da54a2dcd33170067ec409d5a7790c4005a653e516a6e517bc22bde"} Sep 30 14:12:11 crc kubenswrapper[4840]: I0930 14:12:11.191347 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"66a7a4e6-dd49-4589-bd88-2f5fdf0af70f","Type":"ContainerStarted","Data":"932ccc127007755c66f06838f2115e320f6c5102ec86b7d60db6a248d4f4e61a"} Sep 30 14:12:11 crc kubenswrapper[4840]: I0930 14:12:11.191355 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"66a7a4e6-dd49-4589-bd88-2f5fdf0af70f","Type":"ContainerStarted","Data":"1d7dcd053e9c29a5b238c862d154e83c1988fed0daefd5214831ed1adb92cf88"} Sep 30 14:12:12 crc kubenswrapper[4840]: I0930 14:12:12.203590 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"66a7a4e6-dd49-4589-bd88-2f5fdf0af70f","Type":"ContainerStarted","Data":"cebe2792ed1e78f252410b15c1b44d1bf797333b285da822ef83548ca381ec19"} Sep 30 14:12:12 crc kubenswrapper[4840]: I0930 14:12:12.203945 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"66a7a4e6-dd49-4589-bd88-2f5fdf0af70f","Type":"ContainerStarted","Data":"e21e7c98395afb3192462f4e30198c7e7db0134d0fe526f7da68e9eec0422e6a"} Sep 30 14:12:12 crc kubenswrapper[4840]: I0930 14:12:12.245804 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=19.971205563 podStartE2EDuration="26.245779641s" podCreationTimestamp="2025-09-30 14:11:46 +0000 UTC" firstStartedPulling="2025-09-30 14:12:03.828600852 +0000 UTC m=+952.457687275" lastFinishedPulling="2025-09-30 14:12:10.10317493 +0000 UTC m=+958.732261353" observedRunningTime="2025-09-30 14:12:12.234282752 +0000 UTC m=+960.863369195" watchObservedRunningTime="2025-09-30 14:12:12.245779641 +0000 UTC m=+960.874866064" Sep 30 14:12:12 crc kubenswrapper[4840]: I0930 14:12:12.482317 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-k8nq4"] Sep 30 14:12:12 crc kubenswrapper[4840]: I0930 14:12:12.484234 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-k8nq4" Sep 30 14:12:12 crc kubenswrapper[4840]: I0930 14:12:12.490497 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Sep 30 14:12:12 crc kubenswrapper[4840]: I0930 14:12:12.498098 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-k8nq4"] Sep 30 14:12:12 crc kubenswrapper[4840]: I0930 14:12:12.599749 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/16f65bcb-2335-47af-9705-86e0131a8f7c-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-k8nq4\" (UID: \"16f65bcb-2335-47af-9705-86e0131a8f7c\") " pod="openstack/dnsmasq-dns-77585f5f8c-k8nq4" Sep 30 14:12:12 crc kubenswrapper[4840]: I0930 14:12:12.599953 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hnrx\" (UniqueName: \"kubernetes.io/projected/16f65bcb-2335-47af-9705-86e0131a8f7c-kube-api-access-8hnrx\") pod \"dnsmasq-dns-77585f5f8c-k8nq4\" (UID: \"16f65bcb-2335-47af-9705-86e0131a8f7c\") " pod="openstack/dnsmasq-dns-77585f5f8c-k8nq4" Sep 30 14:12:12 crc kubenswrapper[4840]: I0930 14:12:12.600050 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16f65bcb-2335-47af-9705-86e0131a8f7c-config\") pod \"dnsmasq-dns-77585f5f8c-k8nq4\" (UID: \"16f65bcb-2335-47af-9705-86e0131a8f7c\") " pod="openstack/dnsmasq-dns-77585f5f8c-k8nq4" Sep 30 14:12:12 crc kubenswrapper[4840]: I0930 14:12:12.600196 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/16f65bcb-2335-47af-9705-86e0131a8f7c-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-k8nq4\" (UID: \"16f65bcb-2335-47af-9705-86e0131a8f7c\") " pod="openstack/dnsmasq-dns-77585f5f8c-k8nq4" Sep 30 14:12:12 crc kubenswrapper[4840]: I0930 14:12:12.600264 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/16f65bcb-2335-47af-9705-86e0131a8f7c-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-k8nq4\" (UID: \"16f65bcb-2335-47af-9705-86e0131a8f7c\") " pod="openstack/dnsmasq-dns-77585f5f8c-k8nq4" Sep 30 14:12:12 crc kubenswrapper[4840]: I0930 14:12:12.600458 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/16f65bcb-2335-47af-9705-86e0131a8f7c-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-k8nq4\" (UID: \"16f65bcb-2335-47af-9705-86e0131a8f7c\") " pod="openstack/dnsmasq-dns-77585f5f8c-k8nq4" Sep 30 14:12:12 crc kubenswrapper[4840]: I0930 14:12:12.702814 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hnrx\" (UniqueName: \"kubernetes.io/projected/16f65bcb-2335-47af-9705-86e0131a8f7c-kube-api-access-8hnrx\") pod \"dnsmasq-dns-77585f5f8c-k8nq4\" (UID: \"16f65bcb-2335-47af-9705-86e0131a8f7c\") " pod="openstack/dnsmasq-dns-77585f5f8c-k8nq4" Sep 30 14:12:12 crc kubenswrapper[4840]: I0930 14:12:12.702876 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16f65bcb-2335-47af-9705-86e0131a8f7c-config\") pod \"dnsmasq-dns-77585f5f8c-k8nq4\" (UID: \"16f65bcb-2335-47af-9705-86e0131a8f7c\") " pod="openstack/dnsmasq-dns-77585f5f8c-k8nq4" Sep 30 14:12:12 crc kubenswrapper[4840]: I0930 14:12:12.702918 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/16f65bcb-2335-47af-9705-86e0131a8f7c-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-k8nq4\" (UID: \"16f65bcb-2335-47af-9705-86e0131a8f7c\") " pod="openstack/dnsmasq-dns-77585f5f8c-k8nq4" Sep 30 14:12:12 crc kubenswrapper[4840]: I0930 14:12:12.702947 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/16f65bcb-2335-47af-9705-86e0131a8f7c-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-k8nq4\" (UID: \"16f65bcb-2335-47af-9705-86e0131a8f7c\") " pod="openstack/dnsmasq-dns-77585f5f8c-k8nq4" Sep 30 14:12:12 crc kubenswrapper[4840]: I0930 14:12:12.704952 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/16f65bcb-2335-47af-9705-86e0131a8f7c-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-k8nq4\" (UID: \"16f65bcb-2335-47af-9705-86e0131a8f7c\") " pod="openstack/dnsmasq-dns-77585f5f8c-k8nq4" Sep 30 14:12:12 crc kubenswrapper[4840]: I0930 14:12:12.705087 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/16f65bcb-2335-47af-9705-86e0131a8f7c-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-k8nq4\" (UID: \"16f65bcb-2335-47af-9705-86e0131a8f7c\") " pod="openstack/dnsmasq-dns-77585f5f8c-k8nq4" Sep 30 14:12:12 crc kubenswrapper[4840]: I0930 14:12:12.705665 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16f65bcb-2335-47af-9705-86e0131a8f7c-config\") pod \"dnsmasq-dns-77585f5f8c-k8nq4\" (UID: \"16f65bcb-2335-47af-9705-86e0131a8f7c\") " pod="openstack/dnsmasq-dns-77585f5f8c-k8nq4" Sep 30 14:12:12 crc kubenswrapper[4840]: I0930 14:12:12.705697 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/16f65bcb-2335-47af-9705-86e0131a8f7c-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-k8nq4\" (UID: \"16f65bcb-2335-47af-9705-86e0131a8f7c\") " pod="openstack/dnsmasq-dns-77585f5f8c-k8nq4" Sep 30 14:12:12 crc kubenswrapper[4840]: I0930 14:12:12.706355 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/16f65bcb-2335-47af-9705-86e0131a8f7c-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-k8nq4\" (UID: \"16f65bcb-2335-47af-9705-86e0131a8f7c\") " pod="openstack/dnsmasq-dns-77585f5f8c-k8nq4" Sep 30 14:12:12 crc kubenswrapper[4840]: I0930 14:12:12.707030 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/16f65bcb-2335-47af-9705-86e0131a8f7c-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-k8nq4\" (UID: \"16f65bcb-2335-47af-9705-86e0131a8f7c\") " pod="openstack/dnsmasq-dns-77585f5f8c-k8nq4" Sep 30 14:12:12 crc kubenswrapper[4840]: I0930 14:12:12.707111 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/16f65bcb-2335-47af-9705-86e0131a8f7c-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-k8nq4\" (UID: \"16f65bcb-2335-47af-9705-86e0131a8f7c\") " pod="openstack/dnsmasq-dns-77585f5f8c-k8nq4" Sep 30 14:12:12 crc kubenswrapper[4840]: I0930 14:12:12.725981 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hnrx\" (UniqueName: \"kubernetes.io/projected/16f65bcb-2335-47af-9705-86e0131a8f7c-kube-api-access-8hnrx\") pod \"dnsmasq-dns-77585f5f8c-k8nq4\" (UID: \"16f65bcb-2335-47af-9705-86e0131a8f7c\") " pod="openstack/dnsmasq-dns-77585f5f8c-k8nq4" Sep 30 14:12:12 crc kubenswrapper[4840]: I0930 14:12:12.809920 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-k8nq4" Sep 30 14:12:13 crc kubenswrapper[4840]: I0930 14:12:13.269113 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-k8nq4"] Sep 30 14:12:14 crc kubenswrapper[4840]: I0930 14:12:14.235820 4840 generic.go:334] "Generic (PLEG): container finished" podID="16f65bcb-2335-47af-9705-86e0131a8f7c" containerID="cc38a3c8c3fdff136f33150d801d31f8f5cf2972029015e0b6998146246f4c35" exitCode=0 Sep 30 14:12:14 crc kubenswrapper[4840]: I0930 14:12:14.236199 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-k8nq4" event={"ID":"16f65bcb-2335-47af-9705-86e0131a8f7c","Type":"ContainerDied","Data":"cc38a3c8c3fdff136f33150d801d31f8f5cf2972029015e0b6998146246f4c35"} Sep 30 14:12:14 crc kubenswrapper[4840]: I0930 14:12:14.236231 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-k8nq4" event={"ID":"16f65bcb-2335-47af-9705-86e0131a8f7c","Type":"ContainerStarted","Data":"b34ece7c7734d004459e03dac41163ff260ca8475e97f8258b1a1d13671c9f5e"} Sep 30 14:12:15 crc kubenswrapper[4840]: I0930 14:12:15.250263 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-k8nq4" event={"ID":"16f65bcb-2335-47af-9705-86e0131a8f7c","Type":"ContainerStarted","Data":"3e4d847f6209271b5d7feb61da4bc7a0a0745742755f37fd279116bb4f538388"} Sep 30 14:12:15 crc kubenswrapper[4840]: I0930 14:12:15.251126 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77585f5f8c-k8nq4" Sep 30 14:12:15 crc kubenswrapper[4840]: I0930 14:12:15.270790 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77585f5f8c-k8nq4" podStartSLOduration=3.270769806 podStartE2EDuration="3.270769806s" podCreationTimestamp="2025-09-30 14:12:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:12:15.268863376 +0000 UTC m=+963.897949829" watchObservedRunningTime="2025-09-30 14:12:15.270769806 +0000 UTC m=+963.899856229" Sep 30 14:12:20 crc kubenswrapper[4840]: I0930 14:12:20.655773 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:12:20 crc kubenswrapper[4840]: I0930 14:12:20.909976 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Sep 30 14:12:22 crc kubenswrapper[4840]: I0930 14:12:22.430272 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-g9msc"] Sep 30 14:12:22 crc kubenswrapper[4840]: I0930 14:12:22.431507 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-g9msc" Sep 30 14:12:22 crc kubenswrapper[4840]: I0930 14:12:22.452503 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-g9msc"] Sep 30 14:12:22 crc kubenswrapper[4840]: I0930 14:12:22.485049 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smczr\" (UniqueName: \"kubernetes.io/projected/57b18c4f-a2dc-433c-96e0-fbcebd4ecb32-kube-api-access-smczr\") pod \"cinder-db-create-g9msc\" (UID: \"57b18c4f-a2dc-433c-96e0-fbcebd4ecb32\") " pod="openstack/cinder-db-create-g9msc" Sep 30 14:12:22 crc kubenswrapper[4840]: I0930 14:12:22.542906 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-kcv4b"] Sep 30 14:12:22 crc kubenswrapper[4840]: I0930 14:12:22.544130 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-kcv4b" Sep 30 14:12:22 crc kubenswrapper[4840]: I0930 14:12:22.559339 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-kcv4b"] Sep 30 14:12:22 crc kubenswrapper[4840]: I0930 14:12:22.586770 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smczr\" (UniqueName: \"kubernetes.io/projected/57b18c4f-a2dc-433c-96e0-fbcebd4ecb32-kube-api-access-smczr\") pod \"cinder-db-create-g9msc\" (UID: \"57b18c4f-a2dc-433c-96e0-fbcebd4ecb32\") " pod="openstack/cinder-db-create-g9msc" Sep 30 14:12:22 crc kubenswrapper[4840]: I0930 14:12:22.622638 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smczr\" (UniqueName: \"kubernetes.io/projected/57b18c4f-a2dc-433c-96e0-fbcebd4ecb32-kube-api-access-smczr\") pod \"cinder-db-create-g9msc\" (UID: \"57b18c4f-a2dc-433c-96e0-fbcebd4ecb32\") " pod="openstack/cinder-db-create-g9msc" Sep 30 14:12:22 crc kubenswrapper[4840]: I0930 14:12:22.689434 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wj5jc\" (UniqueName: \"kubernetes.io/projected/47747e51-6cb7-4944-a2a1-8011dc25a43a-kube-api-access-wj5jc\") pod \"barbican-db-create-kcv4b\" (UID: \"47747e51-6cb7-4944-a2a1-8011dc25a43a\") " pod="openstack/barbican-db-create-kcv4b" Sep 30 14:12:22 crc kubenswrapper[4840]: I0930 14:12:22.736737 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-92cq6"] Sep 30 14:12:22 crc kubenswrapper[4840]: I0930 14:12:22.738013 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-92cq6" Sep 30 14:12:22 crc kubenswrapper[4840]: I0930 14:12:22.747328 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-92cq6"] Sep 30 14:12:22 crc kubenswrapper[4840]: I0930 14:12:22.756658 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-g9msc" Sep 30 14:12:22 crc kubenswrapper[4840]: I0930 14:12:22.790903 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wj5jc\" (UniqueName: \"kubernetes.io/projected/47747e51-6cb7-4944-a2a1-8011dc25a43a-kube-api-access-wj5jc\") pod \"barbican-db-create-kcv4b\" (UID: \"47747e51-6cb7-4944-a2a1-8011dc25a43a\") " pod="openstack/barbican-db-create-kcv4b" Sep 30 14:12:22 crc kubenswrapper[4840]: I0930 14:12:22.808010 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wj5jc\" (UniqueName: \"kubernetes.io/projected/47747e51-6cb7-4944-a2a1-8011dc25a43a-kube-api-access-wj5jc\") pod \"barbican-db-create-kcv4b\" (UID: \"47747e51-6cb7-4944-a2a1-8011dc25a43a\") " pod="openstack/barbican-db-create-kcv4b" Sep 30 14:12:22 crc kubenswrapper[4840]: I0930 14:12:22.811597 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77585f5f8c-k8nq4" Sep 30 14:12:22 crc kubenswrapper[4840]: I0930 14:12:22.837060 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-xvmrt"] Sep 30 14:12:22 crc kubenswrapper[4840]: I0930 14:12:22.838224 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xvmrt" Sep 30 14:12:22 crc kubenswrapper[4840]: I0930 14:12:22.845377 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Sep 30 14:12:22 crc kubenswrapper[4840]: I0930 14:12:22.845621 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gtvd7" Sep 30 14:12:22 crc kubenswrapper[4840]: I0930 14:12:22.846169 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Sep 30 14:12:22 crc kubenswrapper[4840]: I0930 14:12:22.846279 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Sep 30 14:12:22 crc kubenswrapper[4840]: I0930 14:12:22.859029 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-kcv4b" Sep 30 14:12:22 crc kubenswrapper[4840]: I0930 14:12:22.893297 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5749a5e-2f96-49d9-b1c9-d4b8e231b903-config-data\") pod \"keystone-db-sync-xvmrt\" (UID: \"b5749a5e-2f96-49d9-b1c9-d4b8e231b903\") " pod="openstack/keystone-db-sync-xvmrt" Sep 30 14:12:22 crc kubenswrapper[4840]: I0930 14:12:22.893626 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w76k4\" (UniqueName: \"kubernetes.io/projected/1012193a-00bc-4ffa-81ae-943299c3ac15-kube-api-access-w76k4\") pod \"neutron-db-create-92cq6\" (UID: \"1012193a-00bc-4ffa-81ae-943299c3ac15\") " pod="openstack/neutron-db-create-92cq6" Sep 30 14:12:22 crc kubenswrapper[4840]: I0930 14:12:22.893684 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgdh9\" (UniqueName: \"kubernetes.io/projected/b5749a5e-2f96-49d9-b1c9-d4b8e231b903-kube-api-access-bgdh9\") pod \"keystone-db-sync-xvmrt\" (UID: \"b5749a5e-2f96-49d9-b1c9-d4b8e231b903\") " pod="openstack/keystone-db-sync-xvmrt" Sep 30 14:12:22 crc kubenswrapper[4840]: I0930 14:12:22.893795 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5749a5e-2f96-49d9-b1c9-d4b8e231b903-combined-ca-bundle\") pod \"keystone-db-sync-xvmrt\" (UID: \"b5749a5e-2f96-49d9-b1c9-d4b8e231b903\") " pod="openstack/keystone-db-sync-xvmrt" Sep 30 14:12:22 crc kubenswrapper[4840]: I0930 14:12:22.895485 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-xvmrt"] Sep 30 14:12:22 crc kubenswrapper[4840]: I0930 14:12:22.929467 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-2mqrn"] Sep 30 14:12:22 crc kubenswrapper[4840]: I0930 14:12:22.929750 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-2mqrn" podUID="377cf7b4-32f7-4aa9-85f2-1b62763d9562" containerName="dnsmasq-dns" containerID="cri-o://d93053581c54001e96f6510a1e439aead4da6afe6b19e59c81d8f04be6a78b52" gracePeriod=10 Sep 30 14:12:22 crc kubenswrapper[4840]: I0930 14:12:22.994825 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5749a5e-2f96-49d9-b1c9-d4b8e231b903-combined-ca-bundle\") pod \"keystone-db-sync-xvmrt\" (UID: \"b5749a5e-2f96-49d9-b1c9-d4b8e231b903\") " pod="openstack/keystone-db-sync-xvmrt" Sep 30 14:12:22 crc kubenswrapper[4840]: I0930 14:12:22.994884 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5749a5e-2f96-49d9-b1c9-d4b8e231b903-config-data\") pod \"keystone-db-sync-xvmrt\" (UID: \"b5749a5e-2f96-49d9-b1c9-d4b8e231b903\") " pod="openstack/keystone-db-sync-xvmrt" Sep 30 14:12:22 crc kubenswrapper[4840]: I0930 14:12:22.994911 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w76k4\" (UniqueName: \"kubernetes.io/projected/1012193a-00bc-4ffa-81ae-943299c3ac15-kube-api-access-w76k4\") pod \"neutron-db-create-92cq6\" (UID: \"1012193a-00bc-4ffa-81ae-943299c3ac15\") " pod="openstack/neutron-db-create-92cq6" Sep 30 14:12:22 crc kubenswrapper[4840]: I0930 14:12:22.994965 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgdh9\" (UniqueName: \"kubernetes.io/projected/b5749a5e-2f96-49d9-b1c9-d4b8e231b903-kube-api-access-bgdh9\") pod \"keystone-db-sync-xvmrt\" (UID: \"b5749a5e-2f96-49d9-b1c9-d4b8e231b903\") " pod="openstack/keystone-db-sync-xvmrt" Sep 30 14:12:23 crc kubenswrapper[4840]: I0930 14:12:23.001219 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5749a5e-2f96-49d9-b1c9-d4b8e231b903-combined-ca-bundle\") pod \"keystone-db-sync-xvmrt\" (UID: \"b5749a5e-2f96-49d9-b1c9-d4b8e231b903\") " pod="openstack/keystone-db-sync-xvmrt" Sep 30 14:12:23 crc kubenswrapper[4840]: I0930 14:12:23.002035 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5749a5e-2f96-49d9-b1c9-d4b8e231b903-config-data\") pod \"keystone-db-sync-xvmrt\" (UID: \"b5749a5e-2f96-49d9-b1c9-d4b8e231b903\") " pod="openstack/keystone-db-sync-xvmrt" Sep 30 14:12:23 crc kubenswrapper[4840]: I0930 14:12:23.014766 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgdh9\" (UniqueName: \"kubernetes.io/projected/b5749a5e-2f96-49d9-b1c9-d4b8e231b903-kube-api-access-bgdh9\") pod \"keystone-db-sync-xvmrt\" (UID: \"b5749a5e-2f96-49d9-b1c9-d4b8e231b903\") " pod="openstack/keystone-db-sync-xvmrt" Sep 30 14:12:23 crc kubenswrapper[4840]: I0930 14:12:23.018232 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w76k4\" (UniqueName: \"kubernetes.io/projected/1012193a-00bc-4ffa-81ae-943299c3ac15-kube-api-access-w76k4\") pod \"neutron-db-create-92cq6\" (UID: \"1012193a-00bc-4ffa-81ae-943299c3ac15\") " pod="openstack/neutron-db-create-92cq6" Sep 30 14:12:23 crc kubenswrapper[4840]: I0930 14:12:23.065042 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-92cq6" Sep 30 14:12:23 crc kubenswrapper[4840]: I0930 14:12:23.168640 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-kcv4b"] Sep 30 14:12:23 crc kubenswrapper[4840]: I0930 14:12:23.188835 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xvmrt" Sep 30 14:12:23 crc kubenswrapper[4840]: I0930 14:12:23.324937 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-g9msc"] Sep 30 14:12:23 crc kubenswrapper[4840]: I0930 14:12:23.350214 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-czdht" event={"ID":"a1f15dc9-8926-4c0c-89da-a27216c4cb1a","Type":"ContainerStarted","Data":"2f4453a841d92d974512f9df0a23b410b7163fc5fd3797f2b601df44aa7d14fb"} Sep 30 14:12:23 crc kubenswrapper[4840]: I0930 14:12:23.356983 4840 generic.go:334] "Generic (PLEG): container finished" podID="377cf7b4-32f7-4aa9-85f2-1b62763d9562" containerID="d93053581c54001e96f6510a1e439aead4da6afe6b19e59c81d8f04be6a78b52" exitCode=0 Sep 30 14:12:23 crc kubenswrapper[4840]: I0930 14:12:23.357112 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-2mqrn" event={"ID":"377cf7b4-32f7-4aa9-85f2-1b62763d9562","Type":"ContainerDied","Data":"d93053581c54001e96f6510a1e439aead4da6afe6b19e59c81d8f04be6a78b52"} Sep 30 14:12:23 crc kubenswrapper[4840]: I0930 14:12:23.361227 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-kcv4b" event={"ID":"47747e51-6cb7-4944-a2a1-8011dc25a43a","Type":"ContainerStarted","Data":"f9efaa1e2c4bb5e7aeff21321246dad59ac51be7214ea8bd0331537c9b21129d"} Sep 30 14:12:23 crc kubenswrapper[4840]: I0930 14:12:23.362254 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-g9msc" event={"ID":"57b18c4f-a2dc-433c-96e0-fbcebd4ecb32","Type":"ContainerStarted","Data":"333837ef0e0e66e7068cf4467ea5c96bbe6edcbc3d1737e60eb7975123c92e2b"} Sep 30 14:12:23 crc kubenswrapper[4840]: I0930 14:12:23.395289 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-czdht" podStartSLOduration=3.05322902 podStartE2EDuration="14.39524876s" podCreationTimestamp="2025-09-30 14:12:09 +0000 UTC" firstStartedPulling="2025-09-30 14:12:10.76311088 +0000 UTC m=+959.392197303" lastFinishedPulling="2025-09-30 14:12:22.10513062 +0000 UTC m=+970.734217043" observedRunningTime="2025-09-30 14:12:23.366542942 +0000 UTC m=+971.995629365" watchObservedRunningTime="2025-09-30 14:12:23.39524876 +0000 UTC m=+972.024335183" Sep 30 14:12:23 crc kubenswrapper[4840]: I0930 14:12:23.444225 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-2mqrn" Sep 30 14:12:23 crc kubenswrapper[4840]: I0930 14:12:23.597230 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-92cq6"] Sep 30 14:12:23 crc kubenswrapper[4840]: I0930 14:12:23.614861 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/377cf7b4-32f7-4aa9-85f2-1b62763d9562-ovsdbserver-nb\") pod \"377cf7b4-32f7-4aa9-85f2-1b62763d9562\" (UID: \"377cf7b4-32f7-4aa9-85f2-1b62763d9562\") " Sep 30 14:12:23 crc kubenswrapper[4840]: I0930 14:12:23.614907 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mhmrq\" (UniqueName: \"kubernetes.io/projected/377cf7b4-32f7-4aa9-85f2-1b62763d9562-kube-api-access-mhmrq\") pod \"377cf7b4-32f7-4aa9-85f2-1b62763d9562\" (UID: \"377cf7b4-32f7-4aa9-85f2-1b62763d9562\") " Sep 30 14:12:23 crc kubenswrapper[4840]: I0930 14:12:23.614951 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/377cf7b4-32f7-4aa9-85f2-1b62763d9562-config\") pod \"377cf7b4-32f7-4aa9-85f2-1b62763d9562\" (UID: \"377cf7b4-32f7-4aa9-85f2-1b62763d9562\") " Sep 30 14:12:23 crc kubenswrapper[4840]: I0930 14:12:23.615058 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/377cf7b4-32f7-4aa9-85f2-1b62763d9562-dns-svc\") pod \"377cf7b4-32f7-4aa9-85f2-1b62763d9562\" (UID: \"377cf7b4-32f7-4aa9-85f2-1b62763d9562\") " Sep 30 14:12:23 crc kubenswrapper[4840]: I0930 14:12:23.615099 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/377cf7b4-32f7-4aa9-85f2-1b62763d9562-ovsdbserver-sb\") pod \"377cf7b4-32f7-4aa9-85f2-1b62763d9562\" (UID: \"377cf7b4-32f7-4aa9-85f2-1b62763d9562\") " Sep 30 14:12:23 crc kubenswrapper[4840]: I0930 14:12:23.621021 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/377cf7b4-32f7-4aa9-85f2-1b62763d9562-kube-api-access-mhmrq" (OuterVolumeSpecName: "kube-api-access-mhmrq") pod "377cf7b4-32f7-4aa9-85f2-1b62763d9562" (UID: "377cf7b4-32f7-4aa9-85f2-1b62763d9562"). InnerVolumeSpecName "kube-api-access-mhmrq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:12:23 crc kubenswrapper[4840]: I0930 14:12:23.676754 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/377cf7b4-32f7-4aa9-85f2-1b62763d9562-config" (OuterVolumeSpecName: "config") pod "377cf7b4-32f7-4aa9-85f2-1b62763d9562" (UID: "377cf7b4-32f7-4aa9-85f2-1b62763d9562"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:12:23 crc kubenswrapper[4840]: I0930 14:12:23.678118 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/377cf7b4-32f7-4aa9-85f2-1b62763d9562-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "377cf7b4-32f7-4aa9-85f2-1b62763d9562" (UID: "377cf7b4-32f7-4aa9-85f2-1b62763d9562"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:12:23 crc kubenswrapper[4840]: I0930 14:12:23.680666 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/377cf7b4-32f7-4aa9-85f2-1b62763d9562-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "377cf7b4-32f7-4aa9-85f2-1b62763d9562" (UID: "377cf7b4-32f7-4aa9-85f2-1b62763d9562"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:12:23 crc kubenswrapper[4840]: I0930 14:12:23.689641 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/377cf7b4-32f7-4aa9-85f2-1b62763d9562-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "377cf7b4-32f7-4aa9-85f2-1b62763d9562" (UID: "377cf7b4-32f7-4aa9-85f2-1b62763d9562"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:12:23 crc kubenswrapper[4840]: I0930 14:12:23.717504 4840 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/377cf7b4-32f7-4aa9-85f2-1b62763d9562-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:23 crc kubenswrapper[4840]: I0930 14:12:23.717539 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mhmrq\" (UniqueName: \"kubernetes.io/projected/377cf7b4-32f7-4aa9-85f2-1b62763d9562-kube-api-access-mhmrq\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:23 crc kubenswrapper[4840]: I0930 14:12:23.717568 4840 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/377cf7b4-32f7-4aa9-85f2-1b62763d9562-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:23 crc kubenswrapper[4840]: I0930 14:12:23.717582 4840 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/377cf7b4-32f7-4aa9-85f2-1b62763d9562-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:23 crc kubenswrapper[4840]: I0930 14:12:23.717592 4840 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/377cf7b4-32f7-4aa9-85f2-1b62763d9562-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:23 crc kubenswrapper[4840]: I0930 14:12:23.759249 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-xvmrt"] Sep 30 14:12:23 crc kubenswrapper[4840]: W0930 14:12:23.762415 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5749a5e_2f96_49d9_b1c9_d4b8e231b903.slice/crio-98642218fbbd4bc584891922f591321181b9d3415764f34fc57bd21f02d5ed8d WatchSource:0}: Error finding container 98642218fbbd4bc584891922f591321181b9d3415764f34fc57bd21f02d5ed8d: Status 404 returned error can't find the container with id 98642218fbbd4bc584891922f591321181b9d3415764f34fc57bd21f02d5ed8d Sep 30 14:12:24 crc kubenswrapper[4840]: I0930 14:12:24.371358 4840 generic.go:334] "Generic (PLEG): container finished" podID="57b18c4f-a2dc-433c-96e0-fbcebd4ecb32" containerID="ed388f6da8a23eac58e1d06b8c8262fb4e648b50383158e129180e1f203a3276" exitCode=0 Sep 30 14:12:24 crc kubenswrapper[4840]: I0930 14:12:24.371447 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-g9msc" event={"ID":"57b18c4f-a2dc-433c-96e0-fbcebd4ecb32","Type":"ContainerDied","Data":"ed388f6da8a23eac58e1d06b8c8262fb4e648b50383158e129180e1f203a3276"} Sep 30 14:12:24 crc kubenswrapper[4840]: I0930 14:12:24.375720 4840 generic.go:334] "Generic (PLEG): container finished" podID="1012193a-00bc-4ffa-81ae-943299c3ac15" containerID="7cabc82331616339c77a20dabef3a6e5391ec674437f89c761356362e795e92c" exitCode=0 Sep 30 14:12:24 crc kubenswrapper[4840]: I0930 14:12:24.375933 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-92cq6" event={"ID":"1012193a-00bc-4ffa-81ae-943299c3ac15","Type":"ContainerDied","Data":"7cabc82331616339c77a20dabef3a6e5391ec674437f89c761356362e795e92c"} Sep 30 14:12:24 crc kubenswrapper[4840]: I0930 14:12:24.375970 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-92cq6" event={"ID":"1012193a-00bc-4ffa-81ae-943299c3ac15","Type":"ContainerStarted","Data":"0d238202f6a29dd0eb46954a54289812c35bf698a023058562edf4f2a2b6daa3"} Sep 30 14:12:24 crc kubenswrapper[4840]: I0930 14:12:24.377317 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xvmrt" event={"ID":"b5749a5e-2f96-49d9-b1c9-d4b8e231b903","Type":"ContainerStarted","Data":"98642218fbbd4bc584891922f591321181b9d3415764f34fc57bd21f02d5ed8d"} Sep 30 14:12:24 crc kubenswrapper[4840]: I0930 14:12:24.381669 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-2mqrn" event={"ID":"377cf7b4-32f7-4aa9-85f2-1b62763d9562","Type":"ContainerDied","Data":"93adcd8b5e6c5f908293036f453c8596b3a9519a62b6fe0fa82a852d7d8d25b1"} Sep 30 14:12:24 crc kubenswrapper[4840]: I0930 14:12:24.381690 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-2mqrn" Sep 30 14:12:24 crc kubenswrapper[4840]: I0930 14:12:24.381725 4840 scope.go:117] "RemoveContainer" containerID="d93053581c54001e96f6510a1e439aead4da6afe6b19e59c81d8f04be6a78b52" Sep 30 14:12:24 crc kubenswrapper[4840]: I0930 14:12:24.386846 4840 generic.go:334] "Generic (PLEG): container finished" podID="47747e51-6cb7-4944-a2a1-8011dc25a43a" containerID="a117172c136b13b2d7a0bafe4c4e9077993f24440d861fac8cb69865570a47e9" exitCode=0 Sep 30 14:12:24 crc kubenswrapper[4840]: I0930 14:12:24.387450 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-kcv4b" event={"ID":"47747e51-6cb7-4944-a2a1-8011dc25a43a","Type":"ContainerDied","Data":"a117172c136b13b2d7a0bafe4c4e9077993f24440d861fac8cb69865570a47e9"} Sep 30 14:12:24 crc kubenswrapper[4840]: I0930 14:12:24.404362 4840 scope.go:117] "RemoveContainer" containerID="14af39bb6486cffb0945112230c5f5f02c6f71930e2f48ebc53ca1dc8c694d8a" Sep 30 14:12:24 crc kubenswrapper[4840]: I0930 14:12:24.433668 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-2mqrn"] Sep 30 14:12:24 crc kubenswrapper[4840]: I0930 14:12:24.439205 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-2mqrn"] Sep 30 14:12:25 crc kubenswrapper[4840]: I0930 14:12:25.851243 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-kcv4b" Sep 30 14:12:25 crc kubenswrapper[4840]: I0930 14:12:25.862516 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-g9msc" Sep 30 14:12:25 crc kubenswrapper[4840]: I0930 14:12:25.868192 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-92cq6" Sep 30 14:12:25 crc kubenswrapper[4840]: I0930 14:12:25.948902 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wj5jc\" (UniqueName: \"kubernetes.io/projected/47747e51-6cb7-4944-a2a1-8011dc25a43a-kube-api-access-wj5jc\") pod \"47747e51-6cb7-4944-a2a1-8011dc25a43a\" (UID: \"47747e51-6cb7-4944-a2a1-8011dc25a43a\") " Sep 30 14:12:25 crc kubenswrapper[4840]: I0930 14:12:25.956848 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47747e51-6cb7-4944-a2a1-8011dc25a43a-kube-api-access-wj5jc" (OuterVolumeSpecName: "kube-api-access-wj5jc") pod "47747e51-6cb7-4944-a2a1-8011dc25a43a" (UID: "47747e51-6cb7-4944-a2a1-8011dc25a43a"). InnerVolumeSpecName "kube-api-access-wj5jc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:12:26 crc kubenswrapper[4840]: I0930 14:12:26.050865 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smczr\" (UniqueName: \"kubernetes.io/projected/57b18c4f-a2dc-433c-96e0-fbcebd4ecb32-kube-api-access-smczr\") pod \"57b18c4f-a2dc-433c-96e0-fbcebd4ecb32\" (UID: \"57b18c4f-a2dc-433c-96e0-fbcebd4ecb32\") " Sep 30 14:12:26 crc kubenswrapper[4840]: I0930 14:12:26.051018 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w76k4\" (UniqueName: \"kubernetes.io/projected/1012193a-00bc-4ffa-81ae-943299c3ac15-kube-api-access-w76k4\") pod \"1012193a-00bc-4ffa-81ae-943299c3ac15\" (UID: \"1012193a-00bc-4ffa-81ae-943299c3ac15\") " Sep 30 14:12:26 crc kubenswrapper[4840]: I0930 14:12:26.051479 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wj5jc\" (UniqueName: \"kubernetes.io/projected/47747e51-6cb7-4944-a2a1-8011dc25a43a-kube-api-access-wj5jc\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:26 crc kubenswrapper[4840]: I0930 14:12:26.054637 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57b18c4f-a2dc-433c-96e0-fbcebd4ecb32-kube-api-access-smczr" (OuterVolumeSpecName: "kube-api-access-smczr") pod "57b18c4f-a2dc-433c-96e0-fbcebd4ecb32" (UID: "57b18c4f-a2dc-433c-96e0-fbcebd4ecb32"). InnerVolumeSpecName "kube-api-access-smczr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:12:26 crc kubenswrapper[4840]: I0930 14:12:26.055883 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1012193a-00bc-4ffa-81ae-943299c3ac15-kube-api-access-w76k4" (OuterVolumeSpecName: "kube-api-access-w76k4") pod "1012193a-00bc-4ffa-81ae-943299c3ac15" (UID: "1012193a-00bc-4ffa-81ae-943299c3ac15"). InnerVolumeSpecName "kube-api-access-w76k4". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:12:26 crc kubenswrapper[4840]: I0930 14:12:26.130708 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="377cf7b4-32f7-4aa9-85f2-1b62763d9562" path="/var/lib/kubelet/pods/377cf7b4-32f7-4aa9-85f2-1b62763d9562/volumes" Sep 30 14:12:26 crc kubenswrapper[4840]: I0930 14:12:26.153299 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w76k4\" (UniqueName: \"kubernetes.io/projected/1012193a-00bc-4ffa-81ae-943299c3ac15-kube-api-access-w76k4\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:26 crc kubenswrapper[4840]: I0930 14:12:26.153340 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smczr\" (UniqueName: \"kubernetes.io/projected/57b18c4f-a2dc-433c-96e0-fbcebd4ecb32-kube-api-access-smczr\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:26 crc kubenswrapper[4840]: I0930 14:12:26.416087 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-g9msc" event={"ID":"57b18c4f-a2dc-433c-96e0-fbcebd4ecb32","Type":"ContainerDied","Data":"333837ef0e0e66e7068cf4467ea5c96bbe6edcbc3d1737e60eb7975123c92e2b"} Sep 30 14:12:26 crc kubenswrapper[4840]: I0930 14:12:26.416131 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="333837ef0e0e66e7068cf4467ea5c96bbe6edcbc3d1737e60eb7975123c92e2b" Sep 30 14:12:26 crc kubenswrapper[4840]: I0930 14:12:26.416139 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-g9msc" Sep 30 14:12:26 crc kubenswrapper[4840]: I0930 14:12:26.418639 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-92cq6" event={"ID":"1012193a-00bc-4ffa-81ae-943299c3ac15","Type":"ContainerDied","Data":"0d238202f6a29dd0eb46954a54289812c35bf698a023058562edf4f2a2b6daa3"} Sep 30 14:12:26 crc kubenswrapper[4840]: I0930 14:12:26.419133 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d238202f6a29dd0eb46954a54289812c35bf698a023058562edf4f2a2b6daa3" Sep 30 14:12:26 crc kubenswrapper[4840]: I0930 14:12:26.418667 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-92cq6" Sep 30 14:12:26 crc kubenswrapper[4840]: I0930 14:12:26.424890 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-kcv4b" event={"ID":"47747e51-6cb7-4944-a2a1-8011dc25a43a","Type":"ContainerDied","Data":"f9efaa1e2c4bb5e7aeff21321246dad59ac51be7214ea8bd0331537c9b21129d"} Sep 30 14:12:26 crc kubenswrapper[4840]: I0930 14:12:26.424924 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f9efaa1e2c4bb5e7aeff21321246dad59ac51be7214ea8bd0331537c9b21129d" Sep 30 14:12:26 crc kubenswrapper[4840]: I0930 14:12:26.424976 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-kcv4b" Sep 30 14:12:31 crc kubenswrapper[4840]: I0930 14:12:31.473383 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xvmrt" event={"ID":"b5749a5e-2f96-49d9-b1c9-d4b8e231b903","Type":"ContainerStarted","Data":"0dea90648d1818fd381540b7490dba9e731519612a3256f4d1cda5b6be1ce6fd"} Sep 30 14:12:31 crc kubenswrapper[4840]: I0930 14:12:31.498298 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-xvmrt" podStartSLOduration=2.722921443 podStartE2EDuration="9.498281941s" podCreationTimestamp="2025-09-30 14:12:22 +0000 UTC" firstStartedPulling="2025-09-30 14:12:23.765017323 +0000 UTC m=+972.394103746" lastFinishedPulling="2025-09-30 14:12:30.540377811 +0000 UTC m=+979.169464244" observedRunningTime="2025-09-30 14:12:31.491434653 +0000 UTC m=+980.120521086" watchObservedRunningTime="2025-09-30 14:12:31.498281941 +0000 UTC m=+980.127368364" Sep 30 14:12:32 crc kubenswrapper[4840]: I0930 14:12:32.486679 4840 generic.go:334] "Generic (PLEG): container finished" podID="a1f15dc9-8926-4c0c-89da-a27216c4cb1a" containerID="2f4453a841d92d974512f9df0a23b410b7163fc5fd3797f2b601df44aa7d14fb" exitCode=0 Sep 30 14:12:32 crc kubenswrapper[4840]: I0930 14:12:32.486778 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-czdht" event={"ID":"a1f15dc9-8926-4c0c-89da-a27216c4cb1a","Type":"ContainerDied","Data":"2f4453a841d92d974512f9df0a23b410b7163fc5fd3797f2b601df44aa7d14fb"} Sep 30 14:12:32 crc kubenswrapper[4840]: I0930 14:12:32.494725 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-d231-account-create-2pgzv"] Sep 30 14:12:32 crc kubenswrapper[4840]: E0930 14:12:32.495225 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="377cf7b4-32f7-4aa9-85f2-1b62763d9562" containerName="dnsmasq-dns" Sep 30 14:12:32 crc kubenswrapper[4840]: I0930 14:12:32.495263 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="377cf7b4-32f7-4aa9-85f2-1b62763d9562" containerName="dnsmasq-dns" Sep 30 14:12:32 crc kubenswrapper[4840]: E0930 14:12:32.495297 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47747e51-6cb7-4944-a2a1-8011dc25a43a" containerName="mariadb-database-create" Sep 30 14:12:32 crc kubenswrapper[4840]: I0930 14:12:32.495312 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="47747e51-6cb7-4944-a2a1-8011dc25a43a" containerName="mariadb-database-create" Sep 30 14:12:32 crc kubenswrapper[4840]: E0930 14:12:32.495340 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1012193a-00bc-4ffa-81ae-943299c3ac15" containerName="mariadb-database-create" Sep 30 14:12:32 crc kubenswrapper[4840]: I0930 14:12:32.495353 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="1012193a-00bc-4ffa-81ae-943299c3ac15" containerName="mariadb-database-create" Sep 30 14:12:32 crc kubenswrapper[4840]: E0930 14:12:32.495376 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57b18c4f-a2dc-433c-96e0-fbcebd4ecb32" containerName="mariadb-database-create" Sep 30 14:12:32 crc kubenswrapper[4840]: I0930 14:12:32.495388 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="57b18c4f-a2dc-433c-96e0-fbcebd4ecb32" containerName="mariadb-database-create" Sep 30 14:12:32 crc kubenswrapper[4840]: E0930 14:12:32.495412 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="377cf7b4-32f7-4aa9-85f2-1b62763d9562" containerName="init" Sep 30 14:12:32 crc kubenswrapper[4840]: I0930 14:12:32.495422 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="377cf7b4-32f7-4aa9-85f2-1b62763d9562" containerName="init" Sep 30 14:12:32 crc kubenswrapper[4840]: I0930 14:12:32.495768 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="377cf7b4-32f7-4aa9-85f2-1b62763d9562" containerName="dnsmasq-dns" Sep 30 14:12:32 crc kubenswrapper[4840]: I0930 14:12:32.495806 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="47747e51-6cb7-4944-a2a1-8011dc25a43a" containerName="mariadb-database-create" Sep 30 14:12:32 crc kubenswrapper[4840]: I0930 14:12:32.495824 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="1012193a-00bc-4ffa-81ae-943299c3ac15" containerName="mariadb-database-create" Sep 30 14:12:32 crc kubenswrapper[4840]: I0930 14:12:32.495841 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="57b18c4f-a2dc-433c-96e0-fbcebd4ecb32" containerName="mariadb-database-create" Sep 30 14:12:32 crc kubenswrapper[4840]: I0930 14:12:32.496567 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d231-account-create-2pgzv" Sep 30 14:12:32 crc kubenswrapper[4840]: I0930 14:12:32.499441 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Sep 30 14:12:32 crc kubenswrapper[4840]: I0930 14:12:32.518588 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-d231-account-create-2pgzv"] Sep 30 14:12:32 crc kubenswrapper[4840]: I0930 14:12:32.666698 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8n8mj\" (UniqueName: \"kubernetes.io/projected/fa569ef4-7a5f-42f2-9551-b7e491f0792a-kube-api-access-8n8mj\") pod \"cinder-d231-account-create-2pgzv\" (UID: \"fa569ef4-7a5f-42f2-9551-b7e491f0792a\") " pod="openstack/cinder-d231-account-create-2pgzv" Sep 30 14:12:32 crc kubenswrapper[4840]: I0930 14:12:32.679863 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-aaa0-account-create-fsw9v"] Sep 30 14:12:32 crc kubenswrapper[4840]: I0930 14:12:32.680869 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-aaa0-account-create-fsw9v" Sep 30 14:12:32 crc kubenswrapper[4840]: I0930 14:12:32.687412 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Sep 30 14:12:32 crc kubenswrapper[4840]: I0930 14:12:32.687517 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-aaa0-account-create-fsw9v"] Sep 30 14:12:32 crc kubenswrapper[4840]: I0930 14:12:32.768693 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5cbj\" (UniqueName: \"kubernetes.io/projected/3bafeafc-d531-4c3f-8705-65ce26043333-kube-api-access-f5cbj\") pod \"barbican-aaa0-account-create-fsw9v\" (UID: \"3bafeafc-d531-4c3f-8705-65ce26043333\") " pod="openstack/barbican-aaa0-account-create-fsw9v" Sep 30 14:12:32 crc kubenswrapper[4840]: I0930 14:12:32.768761 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8n8mj\" (UniqueName: \"kubernetes.io/projected/fa569ef4-7a5f-42f2-9551-b7e491f0792a-kube-api-access-8n8mj\") pod \"cinder-d231-account-create-2pgzv\" (UID: \"fa569ef4-7a5f-42f2-9551-b7e491f0792a\") " pod="openstack/cinder-d231-account-create-2pgzv" Sep 30 14:12:32 crc kubenswrapper[4840]: I0930 14:12:32.797137 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8n8mj\" (UniqueName: \"kubernetes.io/projected/fa569ef4-7a5f-42f2-9551-b7e491f0792a-kube-api-access-8n8mj\") pod \"cinder-d231-account-create-2pgzv\" (UID: \"fa569ef4-7a5f-42f2-9551-b7e491f0792a\") " pod="openstack/cinder-d231-account-create-2pgzv" Sep 30 14:12:32 crc kubenswrapper[4840]: I0930 14:12:32.823772 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d231-account-create-2pgzv" Sep 30 14:12:32 crc kubenswrapper[4840]: I0930 14:12:32.870174 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5cbj\" (UniqueName: \"kubernetes.io/projected/3bafeafc-d531-4c3f-8705-65ce26043333-kube-api-access-f5cbj\") pod \"barbican-aaa0-account-create-fsw9v\" (UID: \"3bafeafc-d531-4c3f-8705-65ce26043333\") " pod="openstack/barbican-aaa0-account-create-fsw9v" Sep 30 14:12:32 crc kubenswrapper[4840]: I0930 14:12:32.875797 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-dc26-account-create-69bjd"] Sep 30 14:12:32 crc kubenswrapper[4840]: I0930 14:12:32.877396 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dc26-account-create-69bjd" Sep 30 14:12:32 crc kubenswrapper[4840]: I0930 14:12:32.879585 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Sep 30 14:12:32 crc kubenswrapper[4840]: I0930 14:12:32.887691 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dc26-account-create-69bjd"] Sep 30 14:12:32 crc kubenswrapper[4840]: I0930 14:12:32.902785 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5cbj\" (UniqueName: \"kubernetes.io/projected/3bafeafc-d531-4c3f-8705-65ce26043333-kube-api-access-f5cbj\") pod \"barbican-aaa0-account-create-fsw9v\" (UID: \"3bafeafc-d531-4c3f-8705-65ce26043333\") " pod="openstack/barbican-aaa0-account-create-fsw9v" Sep 30 14:12:32 crc kubenswrapper[4840]: I0930 14:12:32.972375 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9h67z\" (UniqueName: \"kubernetes.io/projected/51a8e57c-0cbf-4072-abb3-696b64c00adb-kube-api-access-9h67z\") pod \"neutron-dc26-account-create-69bjd\" (UID: \"51a8e57c-0cbf-4072-abb3-696b64c00adb\") " pod="openstack/neutron-dc26-account-create-69bjd" Sep 30 14:12:33 crc kubenswrapper[4840]: I0930 14:12:33.003796 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-aaa0-account-create-fsw9v" Sep 30 14:12:33 crc kubenswrapper[4840]: I0930 14:12:33.074041 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9h67z\" (UniqueName: \"kubernetes.io/projected/51a8e57c-0cbf-4072-abb3-696b64c00adb-kube-api-access-9h67z\") pod \"neutron-dc26-account-create-69bjd\" (UID: \"51a8e57c-0cbf-4072-abb3-696b64c00adb\") " pod="openstack/neutron-dc26-account-create-69bjd" Sep 30 14:12:33 crc kubenswrapper[4840]: I0930 14:12:33.096141 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9h67z\" (UniqueName: \"kubernetes.io/projected/51a8e57c-0cbf-4072-abb3-696b64c00adb-kube-api-access-9h67z\") pod \"neutron-dc26-account-create-69bjd\" (UID: \"51a8e57c-0cbf-4072-abb3-696b64c00adb\") " pod="openstack/neutron-dc26-account-create-69bjd" Sep 30 14:12:33 crc kubenswrapper[4840]: I0930 14:12:33.256137 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-d231-account-create-2pgzv"] Sep 30 14:12:33 crc kubenswrapper[4840]: W0930 14:12:33.257836 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfa569ef4_7a5f_42f2_9551_b7e491f0792a.slice/crio-1be2fc078739eddb36294a53be20370fe87eb517efab17de834cf5e48278dbe7 WatchSource:0}: Error finding container 1be2fc078739eddb36294a53be20370fe87eb517efab17de834cf5e48278dbe7: Status 404 returned error can't find the container with id 1be2fc078739eddb36294a53be20370fe87eb517efab17de834cf5e48278dbe7 Sep 30 14:12:33 crc kubenswrapper[4840]: I0930 14:12:33.288709 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dc26-account-create-69bjd" Sep 30 14:12:33 crc kubenswrapper[4840]: I0930 14:12:33.423791 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-aaa0-account-create-fsw9v"] Sep 30 14:12:33 crc kubenswrapper[4840]: W0930 14:12:33.435397 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3bafeafc_d531_4c3f_8705_65ce26043333.slice/crio-2f0c585aded2d965db3b4d42cb7a082eb24f9845cad8b06a7917b41a456209b5 WatchSource:0}: Error finding container 2f0c585aded2d965db3b4d42cb7a082eb24f9845cad8b06a7917b41a456209b5: Status 404 returned error can't find the container with id 2f0c585aded2d965db3b4d42cb7a082eb24f9845cad8b06a7917b41a456209b5 Sep 30 14:12:33 crc kubenswrapper[4840]: I0930 14:12:33.496203 4840 generic.go:334] "Generic (PLEG): container finished" podID="fa569ef4-7a5f-42f2-9551-b7e491f0792a" containerID="55c397fd11aa82dcbaa291c03f23c251a2c0e9edbfb49550400ccd57f18fe70e" exitCode=0 Sep 30 14:12:33 crc kubenswrapper[4840]: I0930 14:12:33.496278 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d231-account-create-2pgzv" event={"ID":"fa569ef4-7a5f-42f2-9551-b7e491f0792a","Type":"ContainerDied","Data":"55c397fd11aa82dcbaa291c03f23c251a2c0e9edbfb49550400ccd57f18fe70e"} Sep 30 14:12:33 crc kubenswrapper[4840]: I0930 14:12:33.496314 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d231-account-create-2pgzv" event={"ID":"fa569ef4-7a5f-42f2-9551-b7e491f0792a","Type":"ContainerStarted","Data":"1be2fc078739eddb36294a53be20370fe87eb517efab17de834cf5e48278dbe7"} Sep 30 14:12:33 crc kubenswrapper[4840]: I0930 14:12:33.497341 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-aaa0-account-create-fsw9v" event={"ID":"3bafeafc-d531-4c3f-8705-65ce26043333","Type":"ContainerStarted","Data":"2f0c585aded2d965db3b4d42cb7a082eb24f9845cad8b06a7917b41a456209b5"} Sep 30 14:12:33 crc kubenswrapper[4840]: I0930 14:12:33.715668 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dc26-account-create-69bjd"] Sep 30 14:12:33 crc kubenswrapper[4840]: I0930 14:12:33.850997 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-czdht" Sep 30 14:12:33 crc kubenswrapper[4840]: I0930 14:12:33.987420 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1f15dc9-8926-4c0c-89da-a27216c4cb1a-config-data\") pod \"a1f15dc9-8926-4c0c-89da-a27216c4cb1a\" (UID: \"a1f15dc9-8926-4c0c-89da-a27216c4cb1a\") " Sep 30 14:12:33 crc kubenswrapper[4840]: I0930 14:12:33.987729 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkk9r\" (UniqueName: \"kubernetes.io/projected/a1f15dc9-8926-4c0c-89da-a27216c4cb1a-kube-api-access-xkk9r\") pod \"a1f15dc9-8926-4c0c-89da-a27216c4cb1a\" (UID: \"a1f15dc9-8926-4c0c-89da-a27216c4cb1a\") " Sep 30 14:12:33 crc kubenswrapper[4840]: I0930 14:12:33.987828 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1f15dc9-8926-4c0c-89da-a27216c4cb1a-combined-ca-bundle\") pod \"a1f15dc9-8926-4c0c-89da-a27216c4cb1a\" (UID: \"a1f15dc9-8926-4c0c-89da-a27216c4cb1a\") " Sep 30 14:12:33 crc kubenswrapper[4840]: I0930 14:12:33.987906 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a1f15dc9-8926-4c0c-89da-a27216c4cb1a-db-sync-config-data\") pod \"a1f15dc9-8926-4c0c-89da-a27216c4cb1a\" (UID: \"a1f15dc9-8926-4c0c-89da-a27216c4cb1a\") " Sep 30 14:12:33 crc kubenswrapper[4840]: I0930 14:12:33.993239 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1f15dc9-8926-4c0c-89da-a27216c4cb1a-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "a1f15dc9-8926-4c0c-89da-a27216c4cb1a" (UID: "a1f15dc9-8926-4c0c-89da-a27216c4cb1a"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:12:33 crc kubenswrapper[4840]: I0930 14:12:33.993299 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1f15dc9-8926-4c0c-89da-a27216c4cb1a-kube-api-access-xkk9r" (OuterVolumeSpecName: "kube-api-access-xkk9r") pod "a1f15dc9-8926-4c0c-89da-a27216c4cb1a" (UID: "a1f15dc9-8926-4c0c-89da-a27216c4cb1a"). InnerVolumeSpecName "kube-api-access-xkk9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:12:34 crc kubenswrapper[4840]: I0930 14:12:34.012767 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1f15dc9-8926-4c0c-89da-a27216c4cb1a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a1f15dc9-8926-4c0c-89da-a27216c4cb1a" (UID: "a1f15dc9-8926-4c0c-89da-a27216c4cb1a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:12:34 crc kubenswrapper[4840]: I0930 14:12:34.028456 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1f15dc9-8926-4c0c-89da-a27216c4cb1a-config-data" (OuterVolumeSpecName: "config-data") pod "a1f15dc9-8926-4c0c-89da-a27216c4cb1a" (UID: "a1f15dc9-8926-4c0c-89da-a27216c4cb1a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:12:34 crc kubenswrapper[4840]: I0930 14:12:34.090173 4840 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a1f15dc9-8926-4c0c-89da-a27216c4cb1a-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:34 crc kubenswrapper[4840]: I0930 14:12:34.090218 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1f15dc9-8926-4c0c-89da-a27216c4cb1a-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:34 crc kubenswrapper[4840]: I0930 14:12:34.090232 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkk9r\" (UniqueName: \"kubernetes.io/projected/a1f15dc9-8926-4c0c-89da-a27216c4cb1a-kube-api-access-xkk9r\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:34 crc kubenswrapper[4840]: I0930 14:12:34.090247 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1f15dc9-8926-4c0c-89da-a27216c4cb1a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:34 crc kubenswrapper[4840]: I0930 14:12:34.506203 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dc26-account-create-69bjd" event={"ID":"51a8e57c-0cbf-4072-abb3-696b64c00adb","Type":"ContainerStarted","Data":"d3c6267fe4fb96f777e003f3fdc08de753fce77bccf131fe0e1b56392dc857cf"} Sep 30 14:12:34 crc kubenswrapper[4840]: I0930 14:12:34.506276 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dc26-account-create-69bjd" event={"ID":"51a8e57c-0cbf-4072-abb3-696b64c00adb","Type":"ContainerStarted","Data":"728eee286b87616538c7fac2028f00de728acb0784c39950486a543d27338e3e"} Sep 30 14:12:34 crc kubenswrapper[4840]: I0930 14:12:34.507944 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-aaa0-account-create-fsw9v" event={"ID":"3bafeafc-d531-4c3f-8705-65ce26043333","Type":"ContainerStarted","Data":"22169f9c22fcb2a49c956e27c3ad963f3388f8869b8a6961e0f3920c2c4784b2"} Sep 30 14:12:34 crc kubenswrapper[4840]: I0930 14:12:34.510329 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-czdht" Sep 30 14:12:34 crc kubenswrapper[4840]: I0930 14:12:34.510351 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-czdht" event={"ID":"a1f15dc9-8926-4c0c-89da-a27216c4cb1a","Type":"ContainerDied","Data":"1b6a70ed22a1c84410ef38d235d326cd24d3a4ec5f6c91f1d4870ebfc878bafc"} Sep 30 14:12:34 crc kubenswrapper[4840]: I0930 14:12:34.510910 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b6a70ed22a1c84410ef38d235d326cd24d3a4ec5f6c91f1d4870ebfc878bafc" Sep 30 14:12:34 crc kubenswrapper[4840]: I0930 14:12:34.526974 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-dc26-account-create-69bjd" podStartSLOduration=2.526935882 podStartE2EDuration="2.526935882s" podCreationTimestamp="2025-09-30 14:12:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:12:34.523472262 +0000 UTC m=+983.152558685" watchObservedRunningTime="2025-09-30 14:12:34.526935882 +0000 UTC m=+983.156022305" Sep 30 14:12:34 crc kubenswrapper[4840]: I0930 14:12:34.556276 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-aaa0-account-create-fsw9v" podStartSLOduration=2.556257827 podStartE2EDuration="2.556257827s" podCreationTimestamp="2025-09-30 14:12:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:12:34.546433721 +0000 UTC m=+983.175520164" watchObservedRunningTime="2025-09-30 14:12:34.556257827 +0000 UTC m=+983.185344250" Sep 30 14:12:35 crc kubenswrapper[4840]: I0930 14:12:34.865546 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-2ts9b"] Sep 30 14:12:35 crc kubenswrapper[4840]: E0930 14:12:34.866370 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1f15dc9-8926-4c0c-89da-a27216c4cb1a" containerName="glance-db-sync" Sep 30 14:12:35 crc kubenswrapper[4840]: I0930 14:12:34.866384 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1f15dc9-8926-4c0c-89da-a27216c4cb1a" containerName="glance-db-sync" Sep 30 14:12:35 crc kubenswrapper[4840]: I0930 14:12:34.867186 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1f15dc9-8926-4c0c-89da-a27216c4cb1a" containerName="glance-db-sync" Sep 30 14:12:35 crc kubenswrapper[4840]: I0930 14:12:34.868011 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-2ts9b" Sep 30 14:12:35 crc kubenswrapper[4840]: I0930 14:12:34.895205 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-2ts9b"] Sep 30 14:12:35 crc kubenswrapper[4840]: I0930 14:12:34.905259 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d231-account-create-2pgzv" Sep 30 14:12:35 crc kubenswrapper[4840]: I0930 14:12:35.002399 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8n8mj\" (UniqueName: \"kubernetes.io/projected/fa569ef4-7a5f-42f2-9551-b7e491f0792a-kube-api-access-8n8mj\") pod \"fa569ef4-7a5f-42f2-9551-b7e491f0792a\" (UID: \"fa569ef4-7a5f-42f2-9551-b7e491f0792a\") " Sep 30 14:12:35 crc kubenswrapper[4840]: I0930 14:12:35.002679 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f130d775-9523-40f2-8ccd-24ab0dedb3b6-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-2ts9b\" (UID: \"f130d775-9523-40f2-8ccd-24ab0dedb3b6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-2ts9b" Sep 30 14:12:35 crc kubenswrapper[4840]: I0930 14:12:35.002751 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f130d775-9523-40f2-8ccd-24ab0dedb3b6-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-2ts9b\" (UID: \"f130d775-9523-40f2-8ccd-24ab0dedb3b6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-2ts9b" Sep 30 14:12:35 crc kubenswrapper[4840]: I0930 14:12:35.002778 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f130d775-9523-40f2-8ccd-24ab0dedb3b6-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-2ts9b\" (UID: \"f130d775-9523-40f2-8ccd-24ab0dedb3b6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-2ts9b" Sep 30 14:12:35 crc kubenswrapper[4840]: I0930 14:12:35.002823 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f130d775-9523-40f2-8ccd-24ab0dedb3b6-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-2ts9b\" (UID: \"f130d775-9523-40f2-8ccd-24ab0dedb3b6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-2ts9b" Sep 30 14:12:35 crc kubenswrapper[4840]: I0930 14:12:35.002876 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f130d775-9523-40f2-8ccd-24ab0dedb3b6-config\") pod \"dnsmasq-dns-7ff5475cc9-2ts9b\" (UID: \"f130d775-9523-40f2-8ccd-24ab0dedb3b6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-2ts9b" Sep 30 14:12:35 crc kubenswrapper[4840]: I0930 14:12:35.002937 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mzw5\" (UniqueName: \"kubernetes.io/projected/f130d775-9523-40f2-8ccd-24ab0dedb3b6-kube-api-access-2mzw5\") pod \"dnsmasq-dns-7ff5475cc9-2ts9b\" (UID: \"f130d775-9523-40f2-8ccd-24ab0dedb3b6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-2ts9b" Sep 30 14:12:35 crc kubenswrapper[4840]: I0930 14:12:35.006817 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa569ef4-7a5f-42f2-9551-b7e491f0792a-kube-api-access-8n8mj" (OuterVolumeSpecName: "kube-api-access-8n8mj") pod "fa569ef4-7a5f-42f2-9551-b7e491f0792a" (UID: "fa569ef4-7a5f-42f2-9551-b7e491f0792a"). InnerVolumeSpecName "kube-api-access-8n8mj". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:12:35 crc kubenswrapper[4840]: I0930 14:12:35.108466 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f130d775-9523-40f2-8ccd-24ab0dedb3b6-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-2ts9b\" (UID: \"f130d775-9523-40f2-8ccd-24ab0dedb3b6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-2ts9b" Sep 30 14:12:35 crc kubenswrapper[4840]: I0930 14:12:35.108515 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f130d775-9523-40f2-8ccd-24ab0dedb3b6-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-2ts9b\" (UID: \"f130d775-9523-40f2-8ccd-24ab0dedb3b6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-2ts9b" Sep 30 14:12:35 crc kubenswrapper[4840]: I0930 14:12:35.108573 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f130d775-9523-40f2-8ccd-24ab0dedb3b6-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-2ts9b\" (UID: \"f130d775-9523-40f2-8ccd-24ab0dedb3b6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-2ts9b" Sep 30 14:12:35 crc kubenswrapper[4840]: I0930 14:12:35.108623 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f130d775-9523-40f2-8ccd-24ab0dedb3b6-config\") pod \"dnsmasq-dns-7ff5475cc9-2ts9b\" (UID: \"f130d775-9523-40f2-8ccd-24ab0dedb3b6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-2ts9b" Sep 30 14:12:35 crc kubenswrapper[4840]: I0930 14:12:35.108670 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mzw5\" (UniqueName: \"kubernetes.io/projected/f130d775-9523-40f2-8ccd-24ab0dedb3b6-kube-api-access-2mzw5\") pod \"dnsmasq-dns-7ff5475cc9-2ts9b\" (UID: \"f130d775-9523-40f2-8ccd-24ab0dedb3b6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-2ts9b" Sep 30 14:12:35 crc kubenswrapper[4840]: I0930 14:12:35.108741 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f130d775-9523-40f2-8ccd-24ab0dedb3b6-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-2ts9b\" (UID: \"f130d775-9523-40f2-8ccd-24ab0dedb3b6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-2ts9b" Sep 30 14:12:35 crc kubenswrapper[4840]: I0930 14:12:35.108804 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8n8mj\" (UniqueName: \"kubernetes.io/projected/fa569ef4-7a5f-42f2-9551-b7e491f0792a-kube-api-access-8n8mj\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:35 crc kubenswrapper[4840]: I0930 14:12:35.110131 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f130d775-9523-40f2-8ccd-24ab0dedb3b6-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-2ts9b\" (UID: \"f130d775-9523-40f2-8ccd-24ab0dedb3b6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-2ts9b" Sep 30 14:12:35 crc kubenswrapper[4840]: I0930 14:12:35.110227 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f130d775-9523-40f2-8ccd-24ab0dedb3b6-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-2ts9b\" (UID: \"f130d775-9523-40f2-8ccd-24ab0dedb3b6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-2ts9b" Sep 30 14:12:35 crc kubenswrapper[4840]: I0930 14:12:35.110227 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f130d775-9523-40f2-8ccd-24ab0dedb3b6-config\") pod \"dnsmasq-dns-7ff5475cc9-2ts9b\" (UID: \"f130d775-9523-40f2-8ccd-24ab0dedb3b6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-2ts9b" Sep 30 14:12:35 crc kubenswrapper[4840]: I0930 14:12:35.110538 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f130d775-9523-40f2-8ccd-24ab0dedb3b6-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-2ts9b\" (UID: \"f130d775-9523-40f2-8ccd-24ab0dedb3b6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-2ts9b" Sep 30 14:12:35 crc kubenswrapper[4840]: I0930 14:12:35.110706 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f130d775-9523-40f2-8ccd-24ab0dedb3b6-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-2ts9b\" (UID: \"f130d775-9523-40f2-8ccd-24ab0dedb3b6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-2ts9b" Sep 30 14:12:35 crc kubenswrapper[4840]: I0930 14:12:35.130781 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mzw5\" (UniqueName: \"kubernetes.io/projected/f130d775-9523-40f2-8ccd-24ab0dedb3b6-kube-api-access-2mzw5\") pod \"dnsmasq-dns-7ff5475cc9-2ts9b\" (UID: \"f130d775-9523-40f2-8ccd-24ab0dedb3b6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-2ts9b" Sep 30 14:12:35 crc kubenswrapper[4840]: I0930 14:12:35.228133 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-2ts9b" Sep 30 14:12:35 crc kubenswrapper[4840]: I0930 14:12:35.518083 4840 generic.go:334] "Generic (PLEG): container finished" podID="51a8e57c-0cbf-4072-abb3-696b64c00adb" containerID="d3c6267fe4fb96f777e003f3fdc08de753fce77bccf131fe0e1b56392dc857cf" exitCode=0 Sep 30 14:12:35 crc kubenswrapper[4840]: I0930 14:12:35.518334 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dc26-account-create-69bjd" event={"ID":"51a8e57c-0cbf-4072-abb3-696b64c00adb","Type":"ContainerDied","Data":"d3c6267fe4fb96f777e003f3fdc08de753fce77bccf131fe0e1b56392dc857cf"} Sep 30 14:12:35 crc kubenswrapper[4840]: I0930 14:12:35.520220 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d231-account-create-2pgzv" event={"ID":"fa569ef4-7a5f-42f2-9551-b7e491f0792a","Type":"ContainerDied","Data":"1be2fc078739eddb36294a53be20370fe87eb517efab17de834cf5e48278dbe7"} Sep 30 14:12:35 crc kubenswrapper[4840]: I0930 14:12:35.520239 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1be2fc078739eddb36294a53be20370fe87eb517efab17de834cf5e48278dbe7" Sep 30 14:12:35 crc kubenswrapper[4840]: I0930 14:12:35.520273 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d231-account-create-2pgzv" Sep 30 14:12:35 crc kubenswrapper[4840]: I0930 14:12:35.523751 4840 generic.go:334] "Generic (PLEG): container finished" podID="3bafeafc-d531-4c3f-8705-65ce26043333" containerID="22169f9c22fcb2a49c956e27c3ad963f3388f8869b8a6961e0f3920c2c4784b2" exitCode=0 Sep 30 14:12:35 crc kubenswrapper[4840]: I0930 14:12:35.523792 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-aaa0-account-create-fsw9v" event={"ID":"3bafeafc-d531-4c3f-8705-65ce26043333","Type":"ContainerDied","Data":"22169f9c22fcb2a49c956e27c3ad963f3388f8869b8a6961e0f3920c2c4784b2"} Sep 30 14:12:35 crc kubenswrapper[4840]: I0930 14:12:35.990838 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-2ts9b"] Sep 30 14:12:36 crc kubenswrapper[4840]: I0930 14:12:36.532045 4840 generic.go:334] "Generic (PLEG): container finished" podID="b5749a5e-2f96-49d9-b1c9-d4b8e231b903" containerID="0dea90648d1818fd381540b7490dba9e731519612a3256f4d1cda5b6be1ce6fd" exitCode=0 Sep 30 14:12:36 crc kubenswrapper[4840]: I0930 14:12:36.532177 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xvmrt" event={"ID":"b5749a5e-2f96-49d9-b1c9-d4b8e231b903","Type":"ContainerDied","Data":"0dea90648d1818fd381540b7490dba9e731519612a3256f4d1cda5b6be1ce6fd"} Sep 30 14:12:36 crc kubenswrapper[4840]: I0930 14:12:36.533831 4840 generic.go:334] "Generic (PLEG): container finished" podID="f130d775-9523-40f2-8ccd-24ab0dedb3b6" containerID="768adc48e802db23e634528323beaae747a1f696476705aa8361f924b2fe2c4d" exitCode=0 Sep 30 14:12:36 crc kubenswrapper[4840]: I0930 14:12:36.533886 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-2ts9b" event={"ID":"f130d775-9523-40f2-8ccd-24ab0dedb3b6","Type":"ContainerDied","Data":"768adc48e802db23e634528323beaae747a1f696476705aa8361f924b2fe2c4d"} Sep 30 14:12:36 crc kubenswrapper[4840]: I0930 14:12:36.533913 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-2ts9b" event={"ID":"f130d775-9523-40f2-8ccd-24ab0dedb3b6","Type":"ContainerStarted","Data":"d087c75baf3c1c057132d4a4fe61bb6dd362c1466e0eeef8e097a866b389dfdf"} Sep 30 14:12:36 crc kubenswrapper[4840]: I0930 14:12:36.761660 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-aaa0-account-create-fsw9v" Sep 30 14:12:36 crc kubenswrapper[4840]: I0930 14:12:36.837913 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dc26-account-create-69bjd" Sep 30 14:12:36 crc kubenswrapper[4840]: I0930 14:12:36.942000 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5cbj\" (UniqueName: \"kubernetes.io/projected/3bafeafc-d531-4c3f-8705-65ce26043333-kube-api-access-f5cbj\") pod \"3bafeafc-d531-4c3f-8705-65ce26043333\" (UID: \"3bafeafc-d531-4c3f-8705-65ce26043333\") " Sep 30 14:12:36 crc kubenswrapper[4840]: I0930 14:12:36.942070 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9h67z\" (UniqueName: \"kubernetes.io/projected/51a8e57c-0cbf-4072-abb3-696b64c00adb-kube-api-access-9h67z\") pod \"51a8e57c-0cbf-4072-abb3-696b64c00adb\" (UID: \"51a8e57c-0cbf-4072-abb3-696b64c00adb\") " Sep 30 14:12:36 crc kubenswrapper[4840]: I0930 14:12:36.947000 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51a8e57c-0cbf-4072-abb3-696b64c00adb-kube-api-access-9h67z" (OuterVolumeSpecName: "kube-api-access-9h67z") pod "51a8e57c-0cbf-4072-abb3-696b64c00adb" (UID: "51a8e57c-0cbf-4072-abb3-696b64c00adb"). InnerVolumeSpecName "kube-api-access-9h67z". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:12:36 crc kubenswrapper[4840]: I0930 14:12:36.947047 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bafeafc-d531-4c3f-8705-65ce26043333-kube-api-access-f5cbj" (OuterVolumeSpecName: "kube-api-access-f5cbj") pod "3bafeafc-d531-4c3f-8705-65ce26043333" (UID: "3bafeafc-d531-4c3f-8705-65ce26043333"). InnerVolumeSpecName "kube-api-access-f5cbj". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:12:37 crc kubenswrapper[4840]: I0930 14:12:37.043642 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5cbj\" (UniqueName: \"kubernetes.io/projected/3bafeafc-d531-4c3f-8705-65ce26043333-kube-api-access-f5cbj\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:37 crc kubenswrapper[4840]: I0930 14:12:37.043671 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9h67z\" (UniqueName: \"kubernetes.io/projected/51a8e57c-0cbf-4072-abb3-696b64c00adb-kube-api-access-9h67z\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:37 crc kubenswrapper[4840]: I0930 14:12:37.543861 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-2ts9b" event={"ID":"f130d775-9523-40f2-8ccd-24ab0dedb3b6","Type":"ContainerStarted","Data":"2e5b6ec6b9d694e4561f6ade7568fdc680fa33bcb4ffdfcd1b2a95fda3dfe0d6"} Sep 30 14:12:37 crc kubenswrapper[4840]: I0930 14:12:37.544087 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7ff5475cc9-2ts9b" Sep 30 14:12:37 crc kubenswrapper[4840]: I0930 14:12:37.545849 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dc26-account-create-69bjd" event={"ID":"51a8e57c-0cbf-4072-abb3-696b64c00adb","Type":"ContainerDied","Data":"728eee286b87616538c7fac2028f00de728acb0784c39950486a543d27338e3e"} Sep 30 14:12:37 crc kubenswrapper[4840]: I0930 14:12:37.545873 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dc26-account-create-69bjd" Sep 30 14:12:37 crc kubenswrapper[4840]: I0930 14:12:37.545889 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="728eee286b87616538c7fac2028f00de728acb0784c39950486a543d27338e3e" Sep 30 14:12:37 crc kubenswrapper[4840]: I0930 14:12:37.548338 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-aaa0-account-create-fsw9v" event={"ID":"3bafeafc-d531-4c3f-8705-65ce26043333","Type":"ContainerDied","Data":"2f0c585aded2d965db3b4d42cb7a082eb24f9845cad8b06a7917b41a456209b5"} Sep 30 14:12:37 crc kubenswrapper[4840]: I0930 14:12:37.548373 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f0c585aded2d965db3b4d42cb7a082eb24f9845cad8b06a7917b41a456209b5" Sep 30 14:12:37 crc kubenswrapper[4840]: I0930 14:12:37.548375 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-aaa0-account-create-fsw9v" Sep 30 14:12:37 crc kubenswrapper[4840]: I0930 14:12:37.582829 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7ff5475cc9-2ts9b" podStartSLOduration=3.582807154 podStartE2EDuration="3.582807154s" podCreationTimestamp="2025-09-30 14:12:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:12:37.580035821 +0000 UTC m=+986.209122264" watchObservedRunningTime="2025-09-30 14:12:37.582807154 +0000 UTC m=+986.211893587" Sep 30 14:12:37 crc kubenswrapper[4840]: I0930 14:12:37.859698 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xvmrt" Sep 30 14:12:37 crc kubenswrapper[4840]: I0930 14:12:37.957221 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5749a5e-2f96-49d9-b1c9-d4b8e231b903-config-data\") pod \"b5749a5e-2f96-49d9-b1c9-d4b8e231b903\" (UID: \"b5749a5e-2f96-49d9-b1c9-d4b8e231b903\") " Sep 30 14:12:37 crc kubenswrapper[4840]: I0930 14:12:37.957304 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5749a5e-2f96-49d9-b1c9-d4b8e231b903-combined-ca-bundle\") pod \"b5749a5e-2f96-49d9-b1c9-d4b8e231b903\" (UID: \"b5749a5e-2f96-49d9-b1c9-d4b8e231b903\") " Sep 30 14:12:37 crc kubenswrapper[4840]: I0930 14:12:37.957413 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgdh9\" (UniqueName: \"kubernetes.io/projected/b5749a5e-2f96-49d9-b1c9-d4b8e231b903-kube-api-access-bgdh9\") pod \"b5749a5e-2f96-49d9-b1c9-d4b8e231b903\" (UID: \"b5749a5e-2f96-49d9-b1c9-d4b8e231b903\") " Sep 30 14:12:37 crc kubenswrapper[4840]: I0930 14:12:37.968023 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5749a5e-2f96-49d9-b1c9-d4b8e231b903-kube-api-access-bgdh9" (OuterVolumeSpecName: "kube-api-access-bgdh9") pod "b5749a5e-2f96-49d9-b1c9-d4b8e231b903" (UID: "b5749a5e-2f96-49d9-b1c9-d4b8e231b903"). InnerVolumeSpecName "kube-api-access-bgdh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:12:37 crc kubenswrapper[4840]: I0930 14:12:37.983539 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5749a5e-2f96-49d9-b1c9-d4b8e231b903-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b5749a5e-2f96-49d9-b1c9-d4b8e231b903" (UID: "b5749a5e-2f96-49d9-b1c9-d4b8e231b903"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:12:38 crc kubenswrapper[4840]: I0930 14:12:38.003971 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5749a5e-2f96-49d9-b1c9-d4b8e231b903-config-data" (OuterVolumeSpecName: "config-data") pod "b5749a5e-2f96-49d9-b1c9-d4b8e231b903" (UID: "b5749a5e-2f96-49d9-b1c9-d4b8e231b903"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:12:38 crc kubenswrapper[4840]: I0930 14:12:38.058847 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgdh9\" (UniqueName: \"kubernetes.io/projected/b5749a5e-2f96-49d9-b1c9-d4b8e231b903-kube-api-access-bgdh9\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:38 crc kubenswrapper[4840]: I0930 14:12:38.058881 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5749a5e-2f96-49d9-b1c9-d4b8e231b903-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:38 crc kubenswrapper[4840]: I0930 14:12:38.058895 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5749a5e-2f96-49d9-b1c9-d4b8e231b903-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:38 crc kubenswrapper[4840]: I0930 14:12:38.563778 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xvmrt" event={"ID":"b5749a5e-2f96-49d9-b1c9-d4b8e231b903","Type":"ContainerDied","Data":"98642218fbbd4bc584891922f591321181b9d3415764f34fc57bd21f02d5ed8d"} Sep 30 14:12:38 crc kubenswrapper[4840]: I0930 14:12:38.564211 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="98642218fbbd4bc584891922f591321181b9d3415764f34fc57bd21f02d5ed8d" Sep 30 14:12:38 crc kubenswrapper[4840]: I0930 14:12:38.563858 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xvmrt" Sep 30 14:12:38 crc kubenswrapper[4840]: I0930 14:12:38.830804 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-2ts9b"] Sep 30 14:12:38 crc kubenswrapper[4840]: I0930 14:12:38.872464 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-pxgwt"] Sep 30 14:12:38 crc kubenswrapper[4840]: E0930 14:12:38.872944 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa569ef4-7a5f-42f2-9551-b7e491f0792a" containerName="mariadb-account-create" Sep 30 14:12:38 crc kubenswrapper[4840]: I0930 14:12:38.872956 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa569ef4-7a5f-42f2-9551-b7e491f0792a" containerName="mariadb-account-create" Sep 30 14:12:38 crc kubenswrapper[4840]: E0930 14:12:38.872969 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5749a5e-2f96-49d9-b1c9-d4b8e231b903" containerName="keystone-db-sync" Sep 30 14:12:38 crc kubenswrapper[4840]: I0930 14:12:38.872975 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5749a5e-2f96-49d9-b1c9-d4b8e231b903" containerName="keystone-db-sync" Sep 30 14:12:38 crc kubenswrapper[4840]: E0930 14:12:38.872985 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51a8e57c-0cbf-4072-abb3-696b64c00adb" containerName="mariadb-account-create" Sep 30 14:12:38 crc kubenswrapper[4840]: I0930 14:12:38.872992 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="51a8e57c-0cbf-4072-abb3-696b64c00adb" containerName="mariadb-account-create" Sep 30 14:12:38 crc kubenswrapper[4840]: E0930 14:12:38.873012 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bafeafc-d531-4c3f-8705-65ce26043333" containerName="mariadb-account-create" Sep 30 14:12:38 crc kubenswrapper[4840]: I0930 14:12:38.873018 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bafeafc-d531-4c3f-8705-65ce26043333" containerName="mariadb-account-create" Sep 30 14:12:38 crc kubenswrapper[4840]: I0930 14:12:38.873159 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa569ef4-7a5f-42f2-9551-b7e491f0792a" containerName="mariadb-account-create" Sep 30 14:12:38 crc kubenswrapper[4840]: I0930 14:12:38.873170 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bafeafc-d531-4c3f-8705-65ce26043333" containerName="mariadb-account-create" Sep 30 14:12:38 crc kubenswrapper[4840]: I0930 14:12:38.873193 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="51a8e57c-0cbf-4072-abb3-696b64c00adb" containerName="mariadb-account-create" Sep 30 14:12:38 crc kubenswrapper[4840]: I0930 14:12:38.873200 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5749a5e-2f96-49d9-b1c9-d4b8e231b903" containerName="keystone-db-sync" Sep 30 14:12:38 crc kubenswrapper[4840]: I0930 14:12:38.876860 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-pxgwt" Sep 30 14:12:38 crc kubenswrapper[4840]: I0930 14:12:38.880755 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Sep 30 14:12:38 crc kubenswrapper[4840]: I0930 14:12:38.881272 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Sep 30 14:12:38 crc kubenswrapper[4840]: I0930 14:12:38.881434 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gtvd7" Sep 30 14:12:38 crc kubenswrapper[4840]: I0930 14:12:38.882228 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-pxgwt"] Sep 30 14:12:38 crc kubenswrapper[4840]: I0930 14:12:38.884956 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Sep 30 14:12:38 crc kubenswrapper[4840]: I0930 14:12:38.901179 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-chrgb"] Sep 30 14:12:38 crc kubenswrapper[4840]: I0930 14:12:38.907324 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-chrgb" Sep 30 14:12:38 crc kubenswrapper[4840]: I0930 14:12:38.919198 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-chrgb"] Sep 30 14:12:38 crc kubenswrapper[4840]: I0930 14:12:38.971867 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1407ce03-21cb-4f1a-9719-6a9aa88a8973-credential-keys\") pod \"keystone-bootstrap-pxgwt\" (UID: \"1407ce03-21cb-4f1a-9719-6a9aa88a8973\") " pod="openstack/keystone-bootstrap-pxgwt" Sep 30 14:12:38 crc kubenswrapper[4840]: I0930 14:12:38.971911 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qr7v9\" (UniqueName: \"kubernetes.io/projected/1407ce03-21cb-4f1a-9719-6a9aa88a8973-kube-api-access-qr7v9\") pod \"keystone-bootstrap-pxgwt\" (UID: \"1407ce03-21cb-4f1a-9719-6a9aa88a8973\") " pod="openstack/keystone-bootstrap-pxgwt" Sep 30 14:12:38 crc kubenswrapper[4840]: I0930 14:12:38.971975 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1407ce03-21cb-4f1a-9719-6a9aa88a8973-combined-ca-bundle\") pod \"keystone-bootstrap-pxgwt\" (UID: \"1407ce03-21cb-4f1a-9719-6a9aa88a8973\") " pod="openstack/keystone-bootstrap-pxgwt" Sep 30 14:12:38 crc kubenswrapper[4840]: I0930 14:12:38.972017 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1407ce03-21cb-4f1a-9719-6a9aa88a8973-config-data\") pod \"keystone-bootstrap-pxgwt\" (UID: \"1407ce03-21cb-4f1a-9719-6a9aa88a8973\") " pod="openstack/keystone-bootstrap-pxgwt" Sep 30 14:12:38 crc kubenswrapper[4840]: I0930 14:12:38.972036 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1407ce03-21cb-4f1a-9719-6a9aa88a8973-fernet-keys\") pod \"keystone-bootstrap-pxgwt\" (UID: \"1407ce03-21cb-4f1a-9719-6a9aa88a8973\") " pod="openstack/keystone-bootstrap-pxgwt" Sep 30 14:12:38 crc kubenswrapper[4840]: I0930 14:12:38.972075 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1407ce03-21cb-4f1a-9719-6a9aa88a8973-scripts\") pod \"keystone-bootstrap-pxgwt\" (UID: \"1407ce03-21cb-4f1a-9719-6a9aa88a8973\") " pod="openstack/keystone-bootstrap-pxgwt" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.071015 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-9vd76"] Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.074388 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48123aa4-9fd6-40d9-9d08-f82a623b068a-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-chrgb\" (UID: \"48123aa4-9fd6-40d9-9d08-f82a623b068a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-chrgb" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.074444 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1407ce03-21cb-4f1a-9719-6a9aa88a8973-combined-ca-bundle\") pod \"keystone-bootstrap-pxgwt\" (UID: \"1407ce03-21cb-4f1a-9719-6a9aa88a8973\") " pod="openstack/keystone-bootstrap-pxgwt" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.074483 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1407ce03-21cb-4f1a-9719-6a9aa88a8973-config-data\") pod \"keystone-bootstrap-pxgwt\" (UID: \"1407ce03-21cb-4f1a-9719-6a9aa88a8973\") " pod="openstack/keystone-bootstrap-pxgwt" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.074502 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1407ce03-21cb-4f1a-9719-6a9aa88a8973-fernet-keys\") pod \"keystone-bootstrap-pxgwt\" (UID: \"1407ce03-21cb-4f1a-9719-6a9aa88a8973\") " pod="openstack/keystone-bootstrap-pxgwt" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.074534 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48123aa4-9fd6-40d9-9d08-f82a623b068a-config\") pod \"dnsmasq-dns-5c5cc7c5ff-chrgb\" (UID: \"48123aa4-9fd6-40d9-9d08-f82a623b068a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-chrgb" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.074567 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1407ce03-21cb-4f1a-9719-6a9aa88a8973-scripts\") pod \"keystone-bootstrap-pxgwt\" (UID: \"1407ce03-21cb-4f1a-9719-6a9aa88a8973\") " pod="openstack/keystone-bootstrap-pxgwt" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.074592 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/48123aa4-9fd6-40d9-9d08-f82a623b068a-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-chrgb\" (UID: \"48123aa4-9fd6-40d9-9d08-f82a623b068a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-chrgb" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.074615 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbbgz\" (UniqueName: \"kubernetes.io/projected/48123aa4-9fd6-40d9-9d08-f82a623b068a-kube-api-access-bbbgz\") pod \"dnsmasq-dns-5c5cc7c5ff-chrgb\" (UID: \"48123aa4-9fd6-40d9-9d08-f82a623b068a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-chrgb" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.074641 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1407ce03-21cb-4f1a-9719-6a9aa88a8973-credential-keys\") pod \"keystone-bootstrap-pxgwt\" (UID: \"1407ce03-21cb-4f1a-9719-6a9aa88a8973\") " pod="openstack/keystone-bootstrap-pxgwt" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.074659 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qr7v9\" (UniqueName: \"kubernetes.io/projected/1407ce03-21cb-4f1a-9719-6a9aa88a8973-kube-api-access-qr7v9\") pod \"keystone-bootstrap-pxgwt\" (UID: \"1407ce03-21cb-4f1a-9719-6a9aa88a8973\") " pod="openstack/keystone-bootstrap-pxgwt" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.074685 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/48123aa4-9fd6-40d9-9d08-f82a623b068a-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-chrgb\" (UID: \"48123aa4-9fd6-40d9-9d08-f82a623b068a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-chrgb" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.074704 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/48123aa4-9fd6-40d9-9d08-f82a623b068a-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-chrgb\" (UID: \"48123aa4-9fd6-40d9-9d08-f82a623b068a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-chrgb" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.081220 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-9vd76" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.081400 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1407ce03-21cb-4f1a-9719-6a9aa88a8973-fernet-keys\") pod \"keystone-bootstrap-pxgwt\" (UID: \"1407ce03-21cb-4f1a-9719-6a9aa88a8973\") " pod="openstack/keystone-bootstrap-pxgwt" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.082576 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1407ce03-21cb-4f1a-9719-6a9aa88a8973-config-data\") pod \"keystone-bootstrap-pxgwt\" (UID: \"1407ce03-21cb-4f1a-9719-6a9aa88a8973\") " pod="openstack/keystone-bootstrap-pxgwt" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.089925 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1407ce03-21cb-4f1a-9719-6a9aa88a8973-scripts\") pod \"keystone-bootstrap-pxgwt\" (UID: \"1407ce03-21cb-4f1a-9719-6a9aa88a8973\") " pod="openstack/keystone-bootstrap-pxgwt" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.091628 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-2qzkw" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.094235 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.094434 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.094858 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-9vd76"] Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.098222 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1407ce03-21cb-4f1a-9719-6a9aa88a8973-credential-keys\") pod \"keystone-bootstrap-pxgwt\" (UID: \"1407ce03-21cb-4f1a-9719-6a9aa88a8973\") " pod="openstack/keystone-bootstrap-pxgwt" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.098288 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1407ce03-21cb-4f1a-9719-6a9aa88a8973-combined-ca-bundle\") pod \"keystone-bootstrap-pxgwt\" (UID: \"1407ce03-21cb-4f1a-9719-6a9aa88a8973\") " pod="openstack/keystone-bootstrap-pxgwt" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.158740 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qr7v9\" (UniqueName: \"kubernetes.io/projected/1407ce03-21cb-4f1a-9719-6a9aa88a8973-kube-api-access-qr7v9\") pod \"keystone-bootstrap-pxgwt\" (UID: \"1407ce03-21cb-4f1a-9719-6a9aa88a8973\") " pod="openstack/keystone-bootstrap-pxgwt" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.163918 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5cc547b489-2p58b"] Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.165366 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5cc547b489-2p58b" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.176673 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/48123aa4-9fd6-40d9-9d08-f82a623b068a-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-chrgb\" (UID: \"48123aa4-9fd6-40d9-9d08-f82a623b068a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-chrgb" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.176734 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbbgz\" (UniqueName: \"kubernetes.io/projected/48123aa4-9fd6-40d9-9d08-f82a623b068a-kube-api-access-bbbgz\") pod \"dnsmasq-dns-5c5cc7c5ff-chrgb\" (UID: \"48123aa4-9fd6-40d9-9d08-f82a623b068a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-chrgb" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.176806 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/242d13bb-6113-421b-83d1-fcccf91a9c68-config-data\") pod \"cinder-db-sync-9vd76\" (UID: \"242d13bb-6113-421b-83d1-fcccf91a9c68\") " pod="openstack/cinder-db-sync-9vd76" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.176831 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/48123aa4-9fd6-40d9-9d08-f82a623b068a-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-chrgb\" (UID: \"48123aa4-9fd6-40d9-9d08-f82a623b068a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-chrgb" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.176853 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/48123aa4-9fd6-40d9-9d08-f82a623b068a-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-chrgb\" (UID: \"48123aa4-9fd6-40d9-9d08-f82a623b068a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-chrgb" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.176909 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/242d13bb-6113-421b-83d1-fcccf91a9c68-db-sync-config-data\") pod \"cinder-db-sync-9vd76\" (UID: \"242d13bb-6113-421b-83d1-fcccf91a9c68\") " pod="openstack/cinder-db-sync-9vd76" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.176946 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48123aa4-9fd6-40d9-9d08-f82a623b068a-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-chrgb\" (UID: \"48123aa4-9fd6-40d9-9d08-f82a623b068a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-chrgb" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.176987 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/242d13bb-6113-421b-83d1-fcccf91a9c68-scripts\") pod \"cinder-db-sync-9vd76\" (UID: \"242d13bb-6113-421b-83d1-fcccf91a9c68\") " pod="openstack/cinder-db-sync-9vd76" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.177023 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjbfr\" (UniqueName: \"kubernetes.io/projected/242d13bb-6113-421b-83d1-fcccf91a9c68-kube-api-access-pjbfr\") pod \"cinder-db-sync-9vd76\" (UID: \"242d13bb-6113-421b-83d1-fcccf91a9c68\") " pod="openstack/cinder-db-sync-9vd76" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.177072 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/242d13bb-6113-421b-83d1-fcccf91a9c68-etc-machine-id\") pod \"cinder-db-sync-9vd76\" (UID: \"242d13bb-6113-421b-83d1-fcccf91a9c68\") " pod="openstack/cinder-db-sync-9vd76" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.177120 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48123aa4-9fd6-40d9-9d08-f82a623b068a-config\") pod \"dnsmasq-dns-5c5cc7c5ff-chrgb\" (UID: \"48123aa4-9fd6-40d9-9d08-f82a623b068a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-chrgb" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.177145 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/242d13bb-6113-421b-83d1-fcccf91a9c68-combined-ca-bundle\") pod \"cinder-db-sync-9vd76\" (UID: \"242d13bb-6113-421b-83d1-fcccf91a9c68\") " pod="openstack/cinder-db-sync-9vd76" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.178566 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/48123aa4-9fd6-40d9-9d08-f82a623b068a-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-chrgb\" (UID: \"48123aa4-9fd6-40d9-9d08-f82a623b068a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-chrgb" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.179572 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/48123aa4-9fd6-40d9-9d08-f82a623b068a-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-chrgb\" (UID: \"48123aa4-9fd6-40d9-9d08-f82a623b068a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-chrgb" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.180225 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/48123aa4-9fd6-40d9-9d08-f82a623b068a-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-chrgb\" (UID: \"48123aa4-9fd6-40d9-9d08-f82a623b068a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-chrgb" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.180797 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.181016 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48123aa4-9fd6-40d9-9d08-f82a623b068a-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-chrgb\" (UID: \"48123aa4-9fd6-40d9-9d08-f82a623b068a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-chrgb" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.181029 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48123aa4-9fd6-40d9-9d08-f82a623b068a-config\") pod \"dnsmasq-dns-5c5cc7c5ff-chrgb\" (UID: \"48123aa4-9fd6-40d9-9d08-f82a623b068a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-chrgb" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.181110 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.181276 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-vpcvp" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.181347 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.199602 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-nplct"] Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.200684 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-nplct" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.202907 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-pxgwt" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.232263 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5cc547b489-2p58b"] Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.232323 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.232477 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-7h4wd" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.232534 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.288947 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/242d13bb-6113-421b-83d1-fcccf91a9c68-db-sync-config-data\") pod \"cinder-db-sync-9vd76\" (UID: \"242d13bb-6113-421b-83d1-fcccf91a9c68\") " pod="openstack/cinder-db-sync-9vd76" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.289005 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a7f0b012-f515-434f-919e-5f0409c8e554-config-data\") pod \"horizon-5cc547b489-2p58b\" (UID: \"a7f0b012-f515-434f-919e-5f0409c8e554\") " pod="openstack/horizon-5cc547b489-2p58b" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.289028 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a7f0b012-f515-434f-919e-5f0409c8e554-horizon-secret-key\") pod \"horizon-5cc547b489-2p58b\" (UID: \"a7f0b012-f515-434f-919e-5f0409c8e554\") " pod="openstack/horizon-5cc547b489-2p58b" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.289049 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/242d13bb-6113-421b-83d1-fcccf91a9c68-scripts\") pod \"cinder-db-sync-9vd76\" (UID: \"242d13bb-6113-421b-83d1-fcccf91a9c68\") " pod="openstack/cinder-db-sync-9vd76" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.289082 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjbfr\" (UniqueName: \"kubernetes.io/projected/242d13bb-6113-421b-83d1-fcccf91a9c68-kube-api-access-pjbfr\") pod \"cinder-db-sync-9vd76\" (UID: \"242d13bb-6113-421b-83d1-fcccf91a9c68\") " pod="openstack/cinder-db-sync-9vd76" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.289112 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/242d13bb-6113-421b-83d1-fcccf91a9c68-etc-machine-id\") pod \"cinder-db-sync-9vd76\" (UID: \"242d13bb-6113-421b-83d1-fcccf91a9c68\") " pod="openstack/cinder-db-sync-9vd76" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.289143 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7f0b012-f515-434f-919e-5f0409c8e554-logs\") pod \"horizon-5cc547b489-2p58b\" (UID: \"a7f0b012-f515-434f-919e-5f0409c8e554\") " pod="openstack/horizon-5cc547b489-2p58b" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.289160 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/242d13bb-6113-421b-83d1-fcccf91a9c68-combined-ca-bundle\") pod \"cinder-db-sync-9vd76\" (UID: \"242d13bb-6113-421b-83d1-fcccf91a9c68\") " pod="openstack/cinder-db-sync-9vd76" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.289200 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a7f0b012-f515-434f-919e-5f0409c8e554-scripts\") pod \"horizon-5cc547b489-2p58b\" (UID: \"a7f0b012-f515-434f-919e-5f0409c8e554\") " pod="openstack/horizon-5cc547b489-2p58b" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.289237 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnxc2\" (UniqueName: \"kubernetes.io/projected/a7f0b012-f515-434f-919e-5f0409c8e554-kube-api-access-hnxc2\") pod \"horizon-5cc547b489-2p58b\" (UID: \"a7f0b012-f515-434f-919e-5f0409c8e554\") " pod="openstack/horizon-5cc547b489-2p58b" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.289265 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/242d13bb-6113-421b-83d1-fcccf91a9c68-config-data\") pod \"cinder-db-sync-9vd76\" (UID: \"242d13bb-6113-421b-83d1-fcccf91a9c68\") " pod="openstack/cinder-db-sync-9vd76" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.290780 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/242d13bb-6113-421b-83d1-fcccf91a9c68-etc-machine-id\") pod \"cinder-db-sync-9vd76\" (UID: \"242d13bb-6113-421b-83d1-fcccf91a9c68\") " pod="openstack/cinder-db-sync-9vd76" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.302566 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.303398 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/242d13bb-6113-421b-83d1-fcccf91a9c68-config-data\") pod \"cinder-db-sync-9vd76\" (UID: \"242d13bb-6113-421b-83d1-fcccf91a9c68\") " pod="openstack/cinder-db-sync-9vd76" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.303894 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/242d13bb-6113-421b-83d1-fcccf91a9c68-db-sync-config-data\") pod \"cinder-db-sync-9vd76\" (UID: \"242d13bb-6113-421b-83d1-fcccf91a9c68\") " pod="openstack/cinder-db-sync-9vd76" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.304436 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.310196 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/242d13bb-6113-421b-83d1-fcccf91a9c68-scripts\") pod \"cinder-db-sync-9vd76\" (UID: \"242d13bb-6113-421b-83d1-fcccf91a9c68\") " pod="openstack/cinder-db-sync-9vd76" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.312517 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/242d13bb-6113-421b-83d1-fcccf91a9c68-combined-ca-bundle\") pod \"cinder-db-sync-9vd76\" (UID: \"242d13bb-6113-421b-83d1-fcccf91a9c68\") " pod="openstack/cinder-db-sync-9vd76" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.312613 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.312837 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.328351 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbbgz\" (UniqueName: \"kubernetes.io/projected/48123aa4-9fd6-40d9-9d08-f82a623b068a-kube-api-access-bbbgz\") pod \"dnsmasq-dns-5c5cc7c5ff-chrgb\" (UID: \"48123aa4-9fd6-40d9-9d08-f82a623b068a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-chrgb" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.352514 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-nplct"] Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.374610 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjbfr\" (UniqueName: \"kubernetes.io/projected/242d13bb-6113-421b-83d1-fcccf91a9c68-kube-api-access-pjbfr\") pod \"cinder-db-sync-9vd76\" (UID: \"242d13bb-6113-421b-83d1-fcccf91a9c68\") " pod="openstack/cinder-db-sync-9vd76" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.376800 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.393030 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7f0b012-f515-434f-919e-5f0409c8e554-logs\") pod \"horizon-5cc547b489-2p58b\" (UID: \"a7f0b012-f515-434f-919e-5f0409c8e554\") " pod="openstack/horizon-5cc547b489-2p58b" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.393082 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/85509efb-abc4-4649-a320-7b9de5487180-config\") pod \"neutron-db-sync-nplct\" (UID: \"85509efb-abc4-4649-a320-7b9de5487180\") " pod="openstack/neutron-db-sync-nplct" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.393120 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a7f0b012-f515-434f-919e-5f0409c8e554-scripts\") pod \"horizon-5cc547b489-2p58b\" (UID: \"a7f0b012-f515-434f-919e-5f0409c8e554\") " pod="openstack/horizon-5cc547b489-2p58b" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.393152 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnxc2\" (UniqueName: \"kubernetes.io/projected/a7f0b012-f515-434f-919e-5f0409c8e554-kube-api-access-hnxc2\") pod \"horizon-5cc547b489-2p58b\" (UID: \"a7f0b012-f515-434f-919e-5f0409c8e554\") " pod="openstack/horizon-5cc547b489-2p58b" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.393179 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85509efb-abc4-4649-a320-7b9de5487180-combined-ca-bundle\") pod \"neutron-db-sync-nplct\" (UID: \"85509efb-abc4-4649-a320-7b9de5487180\") " pod="openstack/neutron-db-sync-nplct" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.393199 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wc5q2\" (UniqueName: \"kubernetes.io/projected/85509efb-abc4-4649-a320-7b9de5487180-kube-api-access-wc5q2\") pod \"neutron-db-sync-nplct\" (UID: \"85509efb-abc4-4649-a320-7b9de5487180\") " pod="openstack/neutron-db-sync-nplct" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.393246 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a7f0b012-f515-434f-919e-5f0409c8e554-config-data\") pod \"horizon-5cc547b489-2p58b\" (UID: \"a7f0b012-f515-434f-919e-5f0409c8e554\") " pod="openstack/horizon-5cc547b489-2p58b" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.393264 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a7f0b012-f515-434f-919e-5f0409c8e554-horizon-secret-key\") pod \"horizon-5cc547b489-2p58b\" (UID: \"a7f0b012-f515-434f-919e-5f0409c8e554\") " pod="openstack/horizon-5cc547b489-2p58b" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.393972 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7f0b012-f515-434f-919e-5f0409c8e554-logs\") pod \"horizon-5cc547b489-2p58b\" (UID: \"a7f0b012-f515-434f-919e-5f0409c8e554\") " pod="openstack/horizon-5cc547b489-2p58b" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.394453 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a7f0b012-f515-434f-919e-5f0409c8e554-scripts\") pod \"horizon-5cc547b489-2p58b\" (UID: \"a7f0b012-f515-434f-919e-5f0409c8e554\") " pod="openstack/horizon-5cc547b489-2p58b" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.395856 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a7f0b012-f515-434f-919e-5f0409c8e554-config-data\") pod \"horizon-5cc547b489-2p58b\" (UID: \"a7f0b012-f515-434f-919e-5f0409c8e554\") " pod="openstack/horizon-5cc547b489-2p58b" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.397936 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-chrgb"] Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.399174 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-chrgb" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.405962 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a7f0b012-f515-434f-919e-5f0409c8e554-horizon-secret-key\") pod \"horizon-5cc547b489-2p58b\" (UID: \"a7f0b012-f515-434f-919e-5f0409c8e554\") " pod="openstack/horizon-5cc547b489-2p58b" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.441613 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-gxf58"] Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.443051 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-gxf58" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.446269 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnxc2\" (UniqueName: \"kubernetes.io/projected/a7f0b012-f515-434f-919e-5f0409c8e554-kube-api-access-hnxc2\") pod \"horizon-5cc547b489-2p58b\" (UID: \"a7f0b012-f515-434f-919e-5f0409c8e554\") " pod="openstack/horizon-5cc547b489-2p58b" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.478892 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7f785c99fc-fpf4g"] Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.480394 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f785c99fc-fpf4g" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.500454 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4a730c4-623d-4ced-a687-bec2d627eae9-scripts\") pod \"ceilometer-0\" (UID: \"e4a730c4-623d-4ced-a687-bec2d627eae9\") " pod="openstack/ceilometer-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.500525 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnb5k\" (UniqueName: \"kubernetes.io/projected/e4a730c4-623d-4ced-a687-bec2d627eae9-kube-api-access-hnb5k\") pod \"ceilometer-0\" (UID: \"e4a730c4-623d-4ced-a687-bec2d627eae9\") " pod="openstack/ceilometer-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.500593 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/85509efb-abc4-4649-a320-7b9de5487180-config\") pod \"neutron-db-sync-nplct\" (UID: \"85509efb-abc4-4649-a320-7b9de5487180\") " pod="openstack/neutron-db-sync-nplct" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.505760 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4a730c4-623d-4ced-a687-bec2d627eae9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e4a730c4-623d-4ced-a687-bec2d627eae9\") " pod="openstack/ceilometer-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.505841 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4a730c4-623d-4ced-a687-bec2d627eae9-config-data\") pod \"ceilometer-0\" (UID: \"e4a730c4-623d-4ced-a687-bec2d627eae9\") " pod="openstack/ceilometer-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.505897 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e4a730c4-623d-4ced-a687-bec2d627eae9-run-httpd\") pod \"ceilometer-0\" (UID: \"e4a730c4-623d-4ced-a687-bec2d627eae9\") " pod="openstack/ceilometer-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.505936 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e4a730c4-623d-4ced-a687-bec2d627eae9-log-httpd\") pod \"ceilometer-0\" (UID: \"e4a730c4-623d-4ced-a687-bec2d627eae9\") " pod="openstack/ceilometer-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.505998 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85509efb-abc4-4649-a320-7b9de5487180-combined-ca-bundle\") pod \"neutron-db-sync-nplct\" (UID: \"85509efb-abc4-4649-a320-7b9de5487180\") " pod="openstack/neutron-db-sync-nplct" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.506021 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wc5q2\" (UniqueName: \"kubernetes.io/projected/85509efb-abc4-4649-a320-7b9de5487180-kube-api-access-wc5q2\") pod \"neutron-db-sync-nplct\" (UID: \"85509efb-abc4-4649-a320-7b9de5487180\") " pod="openstack/neutron-db-sync-nplct" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.506164 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e4a730c4-623d-4ced-a687-bec2d627eae9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e4a730c4-623d-4ced-a687-bec2d627eae9\") " pod="openstack/ceilometer-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.519924 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/85509efb-abc4-4649-a320-7b9de5487180-config\") pod \"neutron-db-sync-nplct\" (UID: \"85509efb-abc4-4649-a320-7b9de5487180\") " pod="openstack/neutron-db-sync-nplct" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.525123 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85509efb-abc4-4649-a320-7b9de5487180-combined-ca-bundle\") pod \"neutron-db-sync-nplct\" (UID: \"85509efb-abc4-4649-a320-7b9de5487180\") " pod="openstack/neutron-db-sync-nplct" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.536618 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-gxf58"] Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.549312 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wc5q2\" (UniqueName: \"kubernetes.io/projected/85509efb-abc4-4649-a320-7b9de5487180-kube-api-access-wc5q2\") pod \"neutron-db-sync-nplct\" (UID: \"85509efb-abc4-4649-a320-7b9de5487180\") " pod="openstack/neutron-db-sync-nplct" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.563732 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-sp7q5"] Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.564882 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-sp7q5" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.570893 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-xq6vq" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.571269 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.572216 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.598058 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7f785c99fc-fpf4g"] Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.603319 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7ff5475cc9-2ts9b" podUID="f130d775-9523-40f2-8ccd-24ab0dedb3b6" containerName="dnsmasq-dns" containerID="cri-o://2e5b6ec6b9d694e4561f6ade7568fdc680fa33bcb4ffdfcd1b2a95fda3dfe0d6" gracePeriod=10 Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.611247 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/47067ffa-fc01-48b4-abcd-4405053eadca-horizon-secret-key\") pod \"horizon-7f785c99fc-fpf4g\" (UID: \"47067ffa-fc01-48b4-abcd-4405053eadca\") " pod="openstack/horizon-7f785c99fc-fpf4g" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.611364 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b73e88ee-cb25-466d-9860-0dd54a1c1f4c-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-gxf58\" (UID: \"b73e88ee-cb25-466d-9860-0dd54a1c1f4c\") " pod="openstack/dnsmasq-dns-8b5c85b87-gxf58" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.611412 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e4a730c4-623d-4ced-a687-bec2d627eae9-log-httpd\") pod \"ceilometer-0\" (UID: \"e4a730c4-623d-4ced-a687-bec2d627eae9\") " pod="openstack/ceilometer-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.611531 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/47067ffa-fc01-48b4-abcd-4405053eadca-scripts\") pod \"horizon-7f785c99fc-fpf4g\" (UID: \"47067ffa-fc01-48b4-abcd-4405053eadca\") " pod="openstack/horizon-7f785c99fc-fpf4g" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.611579 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/47067ffa-fc01-48b4-abcd-4405053eadca-config-data\") pod \"horizon-7f785c99fc-fpf4g\" (UID: \"47067ffa-fc01-48b4-abcd-4405053eadca\") " pod="openstack/horizon-7f785c99fc-fpf4g" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.611623 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9fdp\" (UniqueName: \"kubernetes.io/projected/47067ffa-fc01-48b4-abcd-4405053eadca-kube-api-access-f9fdp\") pod \"horizon-7f785c99fc-fpf4g\" (UID: \"47067ffa-fc01-48b4-abcd-4405053eadca\") " pod="openstack/horizon-7f785c99fc-fpf4g" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.611666 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e4a730c4-623d-4ced-a687-bec2d627eae9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e4a730c4-623d-4ced-a687-bec2d627eae9\") " pod="openstack/ceilometer-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.611736 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4a730c4-623d-4ced-a687-bec2d627eae9-scripts\") pod \"ceilometer-0\" (UID: \"e4a730c4-623d-4ced-a687-bec2d627eae9\") " pod="openstack/ceilometer-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.611759 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b73e88ee-cb25-466d-9860-0dd54a1c1f4c-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-gxf58\" (UID: \"b73e88ee-cb25-466d-9860-0dd54a1c1f4c\") " pod="openstack/dnsmasq-dns-8b5c85b87-gxf58" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.611786 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b73e88ee-cb25-466d-9860-0dd54a1c1f4c-config\") pod \"dnsmasq-dns-8b5c85b87-gxf58\" (UID: \"b73e88ee-cb25-466d-9860-0dd54a1c1f4c\") " pod="openstack/dnsmasq-dns-8b5c85b87-gxf58" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.611847 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnb5k\" (UniqueName: \"kubernetes.io/projected/e4a730c4-623d-4ced-a687-bec2d627eae9-kube-api-access-hnb5k\") pod \"ceilometer-0\" (UID: \"e4a730c4-623d-4ced-a687-bec2d627eae9\") " pod="openstack/ceilometer-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.611898 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4a730c4-623d-4ced-a687-bec2d627eae9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e4a730c4-623d-4ced-a687-bec2d627eae9\") " pod="openstack/ceilometer-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.611989 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b73e88ee-cb25-466d-9860-0dd54a1c1f4c-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-gxf58\" (UID: \"b73e88ee-cb25-466d-9860-0dd54a1c1f4c\") " pod="openstack/dnsmasq-dns-8b5c85b87-gxf58" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.612113 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47067ffa-fc01-48b4-abcd-4405053eadca-logs\") pod \"horizon-7f785c99fc-fpf4g\" (UID: \"47067ffa-fc01-48b4-abcd-4405053eadca\") " pod="openstack/horizon-7f785c99fc-fpf4g" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.612148 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4a730c4-623d-4ced-a687-bec2d627eae9-config-data\") pod \"ceilometer-0\" (UID: \"e4a730c4-623d-4ced-a687-bec2d627eae9\") " pod="openstack/ceilometer-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.612212 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b73e88ee-cb25-466d-9860-0dd54a1c1f4c-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-gxf58\" (UID: \"b73e88ee-cb25-466d-9860-0dd54a1c1f4c\") " pod="openstack/dnsmasq-dns-8b5c85b87-gxf58" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.612230 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkflj\" (UniqueName: \"kubernetes.io/projected/b73e88ee-cb25-466d-9860-0dd54a1c1f4c-kube-api-access-lkflj\") pod \"dnsmasq-dns-8b5c85b87-gxf58\" (UID: \"b73e88ee-cb25-466d-9860-0dd54a1c1f4c\") " pod="openstack/dnsmasq-dns-8b5c85b87-gxf58" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.612294 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e4a730c4-623d-4ced-a687-bec2d627eae9-run-httpd\") pod \"ceilometer-0\" (UID: \"e4a730c4-623d-4ced-a687-bec2d627eae9\") " pod="openstack/ceilometer-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.614232 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e4a730c4-623d-4ced-a687-bec2d627eae9-run-httpd\") pod \"ceilometer-0\" (UID: \"e4a730c4-623d-4ced-a687-bec2d627eae9\") " pod="openstack/ceilometer-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.614389 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-9vd76" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.614866 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e4a730c4-623d-4ced-a687-bec2d627eae9-log-httpd\") pod \"ceilometer-0\" (UID: \"e4a730c4-623d-4ced-a687-bec2d627eae9\") " pod="openstack/ceilometer-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.620857 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4a730c4-623d-4ced-a687-bec2d627eae9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e4a730c4-623d-4ced-a687-bec2d627eae9\") " pod="openstack/ceilometer-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.646177 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e4a730c4-623d-4ced-a687-bec2d627eae9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e4a730c4-623d-4ced-a687-bec2d627eae9\") " pod="openstack/ceilometer-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.655719 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnb5k\" (UniqueName: \"kubernetes.io/projected/e4a730c4-623d-4ced-a687-bec2d627eae9-kube-api-access-hnb5k\") pod \"ceilometer-0\" (UID: \"e4a730c4-623d-4ced-a687-bec2d627eae9\") " pod="openstack/ceilometer-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.655801 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-sp7q5"] Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.656188 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4a730c4-623d-4ced-a687-bec2d627eae9-scripts\") pod \"ceilometer-0\" (UID: \"e4a730c4-623d-4ced-a687-bec2d627eae9\") " pod="openstack/ceilometer-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.663330 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4a730c4-623d-4ced-a687-bec2d627eae9-config-data\") pod \"ceilometer-0\" (UID: \"e4a730c4-623d-4ced-a687-bec2d627eae9\") " pod="openstack/ceilometer-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.671773 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5cc547b489-2p58b" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.673374 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.675048 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.688332 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.688346 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.688710 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.688826 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-7s8kx" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.689699 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-nplct" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.693950 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.707913 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.711236 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.713064 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.715458 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.715670 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.718896 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9fdp\" (UniqueName: \"kubernetes.io/projected/47067ffa-fc01-48b4-abcd-4405053eadca-kube-api-access-f9fdp\") pod \"horizon-7f785c99fc-fpf4g\" (UID: \"47067ffa-fc01-48b4-abcd-4405053eadca\") " pod="openstack/horizon-7f785c99fc-fpf4g" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.719034 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b73e88ee-cb25-466d-9860-0dd54a1c1f4c-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-gxf58\" (UID: \"b73e88ee-cb25-466d-9860-0dd54a1c1f4c\") " pod="openstack/dnsmasq-dns-8b5c85b87-gxf58" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.719069 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b73e88ee-cb25-466d-9860-0dd54a1c1f4c-config\") pod \"dnsmasq-dns-8b5c85b87-gxf58\" (UID: \"b73e88ee-cb25-466d-9860-0dd54a1c1f4c\") " pod="openstack/dnsmasq-dns-8b5c85b87-gxf58" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.719136 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad748b83-4edd-4545-9386-032185c531c5-logs\") pod \"placement-db-sync-sp7q5\" (UID: \"ad748b83-4edd-4545-9386-032185c531c5\") " pod="openstack/placement-db-sync-sp7q5" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.719204 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b73e88ee-cb25-466d-9860-0dd54a1c1f4c-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-gxf58\" (UID: \"b73e88ee-cb25-466d-9860-0dd54a1c1f4c\") " pod="openstack/dnsmasq-dns-8b5c85b87-gxf58" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.719238 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47067ffa-fc01-48b4-abcd-4405053eadca-logs\") pod \"horizon-7f785c99fc-fpf4g\" (UID: \"47067ffa-fc01-48b4-abcd-4405053eadca\") " pod="openstack/horizon-7f785c99fc-fpf4g" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.719280 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ad748b83-4edd-4545-9386-032185c531c5-scripts\") pod \"placement-db-sync-sp7q5\" (UID: \"ad748b83-4edd-4545-9386-032185c531c5\") " pod="openstack/placement-db-sync-sp7q5" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.719302 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad748b83-4edd-4545-9386-032185c531c5-config-data\") pod \"placement-db-sync-sp7q5\" (UID: \"ad748b83-4edd-4545-9386-032185c531c5\") " pod="openstack/placement-db-sync-sp7q5" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.719330 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b73e88ee-cb25-466d-9860-0dd54a1c1f4c-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-gxf58\" (UID: \"b73e88ee-cb25-466d-9860-0dd54a1c1f4c\") " pod="openstack/dnsmasq-dns-8b5c85b87-gxf58" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.719370 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkflj\" (UniqueName: \"kubernetes.io/projected/b73e88ee-cb25-466d-9860-0dd54a1c1f4c-kube-api-access-lkflj\") pod \"dnsmasq-dns-8b5c85b87-gxf58\" (UID: \"b73e88ee-cb25-466d-9860-0dd54a1c1f4c\") " pod="openstack/dnsmasq-dns-8b5c85b87-gxf58" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.719403 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad748b83-4edd-4545-9386-032185c531c5-combined-ca-bundle\") pod \"placement-db-sync-sp7q5\" (UID: \"ad748b83-4edd-4545-9386-032185c531c5\") " pod="openstack/placement-db-sync-sp7q5" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.719459 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/47067ffa-fc01-48b4-abcd-4405053eadca-horizon-secret-key\") pod \"horizon-7f785c99fc-fpf4g\" (UID: \"47067ffa-fc01-48b4-abcd-4405053eadca\") " pod="openstack/horizon-7f785c99fc-fpf4g" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.719505 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b73e88ee-cb25-466d-9860-0dd54a1c1f4c-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-gxf58\" (UID: \"b73e88ee-cb25-466d-9860-0dd54a1c1f4c\") " pod="openstack/dnsmasq-dns-8b5c85b87-gxf58" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.719609 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7g529\" (UniqueName: \"kubernetes.io/projected/ad748b83-4edd-4545-9386-032185c531c5-kube-api-access-7g529\") pod \"placement-db-sync-sp7q5\" (UID: \"ad748b83-4edd-4545-9386-032185c531c5\") " pod="openstack/placement-db-sync-sp7q5" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.719666 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/47067ffa-fc01-48b4-abcd-4405053eadca-scripts\") pod \"horizon-7f785c99fc-fpf4g\" (UID: \"47067ffa-fc01-48b4-abcd-4405053eadca\") " pod="openstack/horizon-7f785c99fc-fpf4g" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.719694 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/47067ffa-fc01-48b4-abcd-4405053eadca-config-data\") pod \"horizon-7f785c99fc-fpf4g\" (UID: \"47067ffa-fc01-48b4-abcd-4405053eadca\") " pod="openstack/horizon-7f785c99fc-fpf4g" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.723835 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b73e88ee-cb25-466d-9860-0dd54a1c1f4c-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-gxf58\" (UID: \"b73e88ee-cb25-466d-9860-0dd54a1c1f4c\") " pod="openstack/dnsmasq-dns-8b5c85b87-gxf58" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.724686 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47067ffa-fc01-48b4-abcd-4405053eadca-logs\") pod \"horizon-7f785c99fc-fpf4g\" (UID: \"47067ffa-fc01-48b4-abcd-4405053eadca\") " pod="openstack/horizon-7f785c99fc-fpf4g" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.726029 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b73e88ee-cb25-466d-9860-0dd54a1c1f4c-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-gxf58\" (UID: \"b73e88ee-cb25-466d-9860-0dd54a1c1f4c\") " pod="openstack/dnsmasq-dns-8b5c85b87-gxf58" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.726442 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b73e88ee-cb25-466d-9860-0dd54a1c1f4c-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-gxf58\" (UID: \"b73e88ee-cb25-466d-9860-0dd54a1c1f4c\") " pod="openstack/dnsmasq-dns-8b5c85b87-gxf58" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.726497 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.727115 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/47067ffa-fc01-48b4-abcd-4405053eadca-scripts\") pod \"horizon-7f785c99fc-fpf4g\" (UID: \"47067ffa-fc01-48b4-abcd-4405053eadca\") " pod="openstack/horizon-7f785c99fc-fpf4g" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.729609 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/47067ffa-fc01-48b4-abcd-4405053eadca-config-data\") pod \"horizon-7f785c99fc-fpf4g\" (UID: \"47067ffa-fc01-48b4-abcd-4405053eadca\") " pod="openstack/horizon-7f785c99fc-fpf4g" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.733612 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b73e88ee-cb25-466d-9860-0dd54a1c1f4c-config\") pod \"dnsmasq-dns-8b5c85b87-gxf58\" (UID: \"b73e88ee-cb25-466d-9860-0dd54a1c1f4c\") " pod="openstack/dnsmasq-dns-8b5c85b87-gxf58" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.733909 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b73e88ee-cb25-466d-9860-0dd54a1c1f4c-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-gxf58\" (UID: \"b73e88ee-cb25-466d-9860-0dd54a1c1f4c\") " pod="openstack/dnsmasq-dns-8b5c85b87-gxf58" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.740106 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/47067ffa-fc01-48b4-abcd-4405053eadca-horizon-secret-key\") pod \"horizon-7f785c99fc-fpf4g\" (UID: \"47067ffa-fc01-48b4-abcd-4405053eadca\") " pod="openstack/horizon-7f785c99fc-fpf4g" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.756428 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkflj\" (UniqueName: \"kubernetes.io/projected/b73e88ee-cb25-466d-9860-0dd54a1c1f4c-kube-api-access-lkflj\") pod \"dnsmasq-dns-8b5c85b87-gxf58\" (UID: \"b73e88ee-cb25-466d-9860-0dd54a1c1f4c\") " pod="openstack/dnsmasq-dns-8b5c85b87-gxf58" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.758755 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9fdp\" (UniqueName: \"kubernetes.io/projected/47067ffa-fc01-48b4-abcd-4405053eadca-kube-api-access-f9fdp\") pod \"horizon-7f785c99fc-fpf4g\" (UID: \"47067ffa-fc01-48b4-abcd-4405053eadca\") " pod="openstack/horizon-7f785c99fc-fpf4g" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.780570 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-gxf58" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.821236 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.821305 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c5d999e-175b-426d-877c-4a698218c363-logs\") pod \"glance-default-external-api-0\" (UID: \"8c5d999e-175b-426d-877c-4a698218c363\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.821331 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c5d999e-175b-426d-877c-4a698218c363-scripts\") pod \"glance-default-external-api-0\" (UID: \"8c5d999e-175b-426d-877c-4a698218c363\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.821360 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad748b83-4edd-4545-9386-032185c531c5-logs\") pod \"placement-db-sync-sp7q5\" (UID: \"ad748b83-4edd-4545-9386-032185c531c5\") " pod="openstack/placement-db-sync-sp7q5" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.821377 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pskrm\" (UniqueName: \"kubernetes.io/projected/8c5d999e-175b-426d-877c-4a698218c363-kube-api-access-pskrm\") pod \"glance-default-external-api-0\" (UID: \"8c5d999e-175b-426d-877c-4a698218c363\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.821407 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.821425 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c5d999e-175b-426d-877c-4a698218c363-config-data\") pod \"glance-default-external-api-0\" (UID: \"8c5d999e-175b-426d-877c-4a698218c363\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.821441 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.821461 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.821477 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4bfb\" (UniqueName: \"kubernetes.io/projected/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-kube-api-access-f4bfb\") pod \"glance-default-internal-api-0\" (UID: \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.821495 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"8c5d999e-175b-426d-877c-4a698218c363\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.821518 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.821537 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ad748b83-4edd-4545-9386-032185c531c5-scripts\") pod \"placement-db-sync-sp7q5\" (UID: \"ad748b83-4edd-4545-9386-032185c531c5\") " pod="openstack/placement-db-sync-sp7q5" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.821567 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad748b83-4edd-4545-9386-032185c531c5-config-data\") pod \"placement-db-sync-sp7q5\" (UID: \"ad748b83-4edd-4545-9386-032185c531c5\") " pod="openstack/placement-db-sync-sp7q5" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.821597 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad748b83-4edd-4545-9386-032185c531c5-combined-ca-bundle\") pod \"placement-db-sync-sp7q5\" (UID: \"ad748b83-4edd-4545-9386-032185c531c5\") " pod="openstack/placement-db-sync-sp7q5" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.821617 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.821635 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-logs\") pod \"glance-default-internal-api-0\" (UID: \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.821649 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c5d999e-175b-426d-877c-4a698218c363-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8c5d999e-175b-426d-877c-4a698218c363\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.821677 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8c5d999e-175b-426d-877c-4a698218c363-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8c5d999e-175b-426d-877c-4a698218c363\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.821699 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c5d999e-175b-426d-877c-4a698218c363-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8c5d999e-175b-426d-877c-4a698218c363\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.821721 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7g529\" (UniqueName: \"kubernetes.io/projected/ad748b83-4edd-4545-9386-032185c531c5-kube-api-access-7g529\") pod \"placement-db-sync-sp7q5\" (UID: \"ad748b83-4edd-4545-9386-032185c531c5\") " pod="openstack/placement-db-sync-sp7q5" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.822339 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad748b83-4edd-4545-9386-032185c531c5-logs\") pod \"placement-db-sync-sp7q5\" (UID: \"ad748b83-4edd-4545-9386-032185c531c5\") " pod="openstack/placement-db-sync-sp7q5" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.827027 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ad748b83-4edd-4545-9386-032185c531c5-scripts\") pod \"placement-db-sync-sp7q5\" (UID: \"ad748b83-4edd-4545-9386-032185c531c5\") " pod="openstack/placement-db-sync-sp7q5" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.831106 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad748b83-4edd-4545-9386-032185c531c5-config-data\") pod \"placement-db-sync-sp7q5\" (UID: \"ad748b83-4edd-4545-9386-032185c531c5\") " pod="openstack/placement-db-sync-sp7q5" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.832396 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad748b83-4edd-4545-9386-032185c531c5-combined-ca-bundle\") pod \"placement-db-sync-sp7q5\" (UID: \"ad748b83-4edd-4545-9386-032185c531c5\") " pod="openstack/placement-db-sync-sp7q5" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.835030 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f785c99fc-fpf4g" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.841136 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7g529\" (UniqueName: \"kubernetes.io/projected/ad748b83-4edd-4545-9386-032185c531c5-kube-api-access-7g529\") pod \"placement-db-sync-sp7q5\" (UID: \"ad748b83-4edd-4545-9386-032185c531c5\") " pod="openstack/placement-db-sync-sp7q5" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.906281 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-sp7q5" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.922676 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c5d999e-175b-426d-877c-4a698218c363-logs\") pod \"glance-default-external-api-0\" (UID: \"8c5d999e-175b-426d-877c-4a698218c363\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.922724 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c5d999e-175b-426d-877c-4a698218c363-scripts\") pod \"glance-default-external-api-0\" (UID: \"8c5d999e-175b-426d-877c-4a698218c363\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.922757 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pskrm\" (UniqueName: \"kubernetes.io/projected/8c5d999e-175b-426d-877c-4a698218c363-kube-api-access-pskrm\") pod \"glance-default-external-api-0\" (UID: \"8c5d999e-175b-426d-877c-4a698218c363\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.922779 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.922799 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c5d999e-175b-426d-877c-4a698218c363-config-data\") pod \"glance-default-external-api-0\" (UID: \"8c5d999e-175b-426d-877c-4a698218c363\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.922815 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.922832 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.922848 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4bfb\" (UniqueName: \"kubernetes.io/projected/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-kube-api-access-f4bfb\") pod \"glance-default-internal-api-0\" (UID: \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.922868 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"8c5d999e-175b-426d-877c-4a698218c363\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.922889 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.922920 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.922940 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-logs\") pod \"glance-default-internal-api-0\" (UID: \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.922954 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c5d999e-175b-426d-877c-4a698218c363-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8c5d999e-175b-426d-877c-4a698218c363\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.922979 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8c5d999e-175b-426d-877c-4a698218c363-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8c5d999e-175b-426d-877c-4a698218c363\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.923000 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c5d999e-175b-426d-877c-4a698218c363-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8c5d999e-175b-426d-877c-4a698218c363\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.923037 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.923460 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c5d999e-175b-426d-877c-4a698218c363-logs\") pod \"glance-default-external-api-0\" (UID: \"8c5d999e-175b-426d-877c-4a698218c363\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.923825 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-logs\") pod \"glance-default-internal-api-0\" (UID: \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.924071 4840 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"8c5d999e-175b-426d-877c-4a698218c363\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.924529 4840 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-internal-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.926334 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8c5d999e-175b-426d-877c-4a698218c363-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8c5d999e-175b-426d-877c-4a698218c363\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.928138 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.934170 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.939136 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c5d999e-175b-426d-877c-4a698218c363-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8c5d999e-175b-426d-877c-4a698218c363\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.941432 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.945477 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-pxgwt"] Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.945648 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c5d999e-175b-426d-877c-4a698218c363-scripts\") pod \"glance-default-external-api-0\" (UID: \"8c5d999e-175b-426d-877c-4a698218c363\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.946535 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c5d999e-175b-426d-877c-4a698218c363-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8c5d999e-175b-426d-877c-4a698218c363\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.946760 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pskrm\" (UniqueName: \"kubernetes.io/projected/8c5d999e-175b-426d-877c-4a698218c363-kube-api-access-pskrm\") pod \"glance-default-external-api-0\" (UID: \"8c5d999e-175b-426d-877c-4a698218c363\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.947584 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c5d999e-175b-426d-877c-4a698218c363-config-data\") pod \"glance-default-external-api-0\" (UID: \"8c5d999e-175b-426d-877c-4a698218c363\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.947857 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.948386 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4bfb\" (UniqueName: \"kubernetes.io/projected/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-kube-api-access-f4bfb\") pod \"glance-default-internal-api-0\" (UID: \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.949989 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.966313 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:39 crc kubenswrapper[4840]: I0930 14:12:39.970098 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"8c5d999e-175b-426d-877c-4a698218c363\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.018428 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.062207 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.098215 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-chrgb"] Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.394077 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5cc547b489-2p58b"] Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.463966 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-2ts9b" Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.540263 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f130d775-9523-40f2-8ccd-24ab0dedb3b6-ovsdbserver-sb\") pod \"f130d775-9523-40f2-8ccd-24ab0dedb3b6\" (UID: \"f130d775-9523-40f2-8ccd-24ab0dedb3b6\") " Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.540317 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f130d775-9523-40f2-8ccd-24ab0dedb3b6-dns-swift-storage-0\") pod \"f130d775-9523-40f2-8ccd-24ab0dedb3b6\" (UID: \"f130d775-9523-40f2-8ccd-24ab0dedb3b6\") " Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.540356 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f130d775-9523-40f2-8ccd-24ab0dedb3b6-ovsdbserver-nb\") pod \"f130d775-9523-40f2-8ccd-24ab0dedb3b6\" (UID: \"f130d775-9523-40f2-8ccd-24ab0dedb3b6\") " Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.540393 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f130d775-9523-40f2-8ccd-24ab0dedb3b6-dns-svc\") pod \"f130d775-9523-40f2-8ccd-24ab0dedb3b6\" (UID: \"f130d775-9523-40f2-8ccd-24ab0dedb3b6\") " Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.540467 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f130d775-9523-40f2-8ccd-24ab0dedb3b6-config\") pod \"f130d775-9523-40f2-8ccd-24ab0dedb3b6\" (UID: \"f130d775-9523-40f2-8ccd-24ab0dedb3b6\") " Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.540612 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mzw5\" (UniqueName: \"kubernetes.io/projected/f130d775-9523-40f2-8ccd-24ab0dedb3b6-kube-api-access-2mzw5\") pod \"f130d775-9523-40f2-8ccd-24ab0dedb3b6\" (UID: \"f130d775-9523-40f2-8ccd-24ab0dedb3b6\") " Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.545045 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f130d775-9523-40f2-8ccd-24ab0dedb3b6-kube-api-access-2mzw5" (OuterVolumeSpecName: "kube-api-access-2mzw5") pod "f130d775-9523-40f2-8ccd-24ab0dedb3b6" (UID: "f130d775-9523-40f2-8ccd-24ab0dedb3b6"). InnerVolumeSpecName "kube-api-access-2mzw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.564835 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-nplct"] Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.576302 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-9vd76"] Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.608114 4840 generic.go:334] "Generic (PLEG): container finished" podID="48123aa4-9fd6-40d9-9d08-f82a623b068a" containerID="7b5e16fbefb4004f06014a8628f8180f0b562df2c673b5dde92dbbadff996f16" exitCode=0 Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.608331 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-chrgb" event={"ID":"48123aa4-9fd6-40d9-9d08-f82a623b068a","Type":"ContainerDied","Data":"7b5e16fbefb4004f06014a8628f8180f0b562df2c673b5dde92dbbadff996f16"} Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.608363 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-chrgb" event={"ID":"48123aa4-9fd6-40d9-9d08-f82a623b068a","Type":"ContainerStarted","Data":"631461d277dde56e8837ec1d908468f4d572fac6cb30373c8e69c807cdbc02ca"} Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.615798 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-nplct" event={"ID":"85509efb-abc4-4649-a320-7b9de5487180","Type":"ContainerStarted","Data":"23dcba2ba3c0ac8d8de12a191bf04af09ebebac19e5a6356c0486cd7fc9d4101"} Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.624967 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-pxgwt" event={"ID":"1407ce03-21cb-4f1a-9719-6a9aa88a8973","Type":"ContainerStarted","Data":"9f12e56d9a170038e99c98ffb6f13331191d2fdd43eecac910f50328d12487f7"} Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.625014 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-pxgwt" event={"ID":"1407ce03-21cb-4f1a-9719-6a9aa88a8973","Type":"ContainerStarted","Data":"4ddbbee15575c768bf713c18158e98db8ba254d7abc083844f81c3122569692c"} Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.627353 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f130d775-9523-40f2-8ccd-24ab0dedb3b6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f130d775-9523-40f2-8ccd-24ab0dedb3b6" (UID: "f130d775-9523-40f2-8ccd-24ab0dedb3b6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.647646 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2mzw5\" (UniqueName: \"kubernetes.io/projected/f130d775-9523-40f2-8ccd-24ab0dedb3b6-kube-api-access-2mzw5\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.647751 4840 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f130d775-9523-40f2-8ccd-24ab0dedb3b6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.648442 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5cc547b489-2p58b" event={"ID":"a7f0b012-f515-434f-919e-5f0409c8e554","Type":"ContainerStarted","Data":"a63a10d7ebc945cd9f18a75e8c7429c4f157b2e5fdb85475b333d02a5f7fce57"} Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.651238 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7f785c99fc-fpf4g"] Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.677297 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-pxgwt" podStartSLOduration=2.6772769309999997 podStartE2EDuration="2.677276931s" podCreationTimestamp="2025-09-30 14:12:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:12:40.674976541 +0000 UTC m=+989.304062964" watchObservedRunningTime="2025-09-30 14:12:40.677276931 +0000 UTC m=+989.306363354" Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.680116 4840 generic.go:334] "Generic (PLEG): container finished" podID="f130d775-9523-40f2-8ccd-24ab0dedb3b6" containerID="2e5b6ec6b9d694e4561f6ade7568fdc680fa33bcb4ffdfcd1b2a95fda3dfe0d6" exitCode=0 Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.680212 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-2ts9b" event={"ID":"f130d775-9523-40f2-8ccd-24ab0dedb3b6","Type":"ContainerDied","Data":"2e5b6ec6b9d694e4561f6ade7568fdc680fa33bcb4ffdfcd1b2a95fda3dfe0d6"} Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.681282 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-2ts9b" event={"ID":"f130d775-9523-40f2-8ccd-24ab0dedb3b6","Type":"ContainerDied","Data":"d087c75baf3c1c057132d4a4fe61bb6dd362c1466e0eeef8e097a866b389dfdf"} Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.681322 4840 scope.go:117] "RemoveContainer" containerID="2e5b6ec6b9d694e4561f6ade7568fdc680fa33bcb4ffdfcd1b2a95fda3dfe0d6" Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.681929 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-2ts9b" Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.684795 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f130d775-9523-40f2-8ccd-24ab0dedb3b6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f130d775-9523-40f2-8ccd-24ab0dedb3b6" (UID: "f130d775-9523-40f2-8ccd-24ab0dedb3b6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.691007 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f130d775-9523-40f2-8ccd-24ab0dedb3b6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f130d775-9523-40f2-8ccd-24ab0dedb3b6" (UID: "f130d775-9523-40f2-8ccd-24ab0dedb3b6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.697637 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.702249 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f130d775-9523-40f2-8ccd-24ab0dedb3b6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f130d775-9523-40f2-8ccd-24ab0dedb3b6" (UID: "f130d775-9523-40f2-8ccd-24ab0dedb3b6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.704619 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f130d775-9523-40f2-8ccd-24ab0dedb3b6-config" (OuterVolumeSpecName: "config") pod "f130d775-9523-40f2-8ccd-24ab0dedb3b6" (UID: "f130d775-9523-40f2-8ccd-24ab0dedb3b6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.729251 4840 scope.go:117] "RemoveContainer" containerID="768adc48e802db23e634528323beaae747a1f696476705aa8361f924b2fe2c4d" Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.738866 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-gxf58"] Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.760074 4840 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f130d775-9523-40f2-8ccd-24ab0dedb3b6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.760103 4840 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f130d775-9523-40f2-8ccd-24ab0dedb3b6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.760114 4840 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f130d775-9523-40f2-8ccd-24ab0dedb3b6-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.760127 4840 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f130d775-9523-40f2-8ccd-24ab0dedb3b6-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.793865 4840 scope.go:117] "RemoveContainer" containerID="2e5b6ec6b9d694e4561f6ade7568fdc680fa33bcb4ffdfcd1b2a95fda3dfe0d6" Sep 30 14:12:40 crc kubenswrapper[4840]: E0930 14:12:40.795334 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e5b6ec6b9d694e4561f6ade7568fdc680fa33bcb4ffdfcd1b2a95fda3dfe0d6\": container with ID starting with 2e5b6ec6b9d694e4561f6ade7568fdc680fa33bcb4ffdfcd1b2a95fda3dfe0d6 not found: ID does not exist" containerID="2e5b6ec6b9d694e4561f6ade7568fdc680fa33bcb4ffdfcd1b2a95fda3dfe0d6" Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.795360 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e5b6ec6b9d694e4561f6ade7568fdc680fa33bcb4ffdfcd1b2a95fda3dfe0d6"} err="failed to get container status \"2e5b6ec6b9d694e4561f6ade7568fdc680fa33bcb4ffdfcd1b2a95fda3dfe0d6\": rpc error: code = NotFound desc = could not find container \"2e5b6ec6b9d694e4561f6ade7568fdc680fa33bcb4ffdfcd1b2a95fda3dfe0d6\": container with ID starting with 2e5b6ec6b9d694e4561f6ade7568fdc680fa33bcb4ffdfcd1b2a95fda3dfe0d6 not found: ID does not exist" Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.795382 4840 scope.go:117] "RemoveContainer" containerID="768adc48e802db23e634528323beaae747a1f696476705aa8361f924b2fe2c4d" Sep 30 14:12:40 crc kubenswrapper[4840]: E0930 14:12:40.795808 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"768adc48e802db23e634528323beaae747a1f696476705aa8361f924b2fe2c4d\": container with ID starting with 768adc48e802db23e634528323beaae747a1f696476705aa8361f924b2fe2c4d not found: ID does not exist" containerID="768adc48e802db23e634528323beaae747a1f696476705aa8361f924b2fe2c4d" Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.795828 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"768adc48e802db23e634528323beaae747a1f696476705aa8361f924b2fe2c4d"} err="failed to get container status \"768adc48e802db23e634528323beaae747a1f696476705aa8361f924b2fe2c4d\": rpc error: code = NotFound desc = could not find container \"768adc48e802db23e634528323beaae747a1f696476705aa8361f924b2fe2c4d\": container with ID starting with 768adc48e802db23e634528323beaae747a1f696476705aa8361f924b2fe2c4d not found: ID does not exist" Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.889776 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-sp7q5"] Sep 30 14:12:40 crc kubenswrapper[4840]: I0930 14:12:40.967257 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-chrgb" Sep 30 14:12:41 crc kubenswrapper[4840]: I0930 14:12:41.045212 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-2ts9b"] Sep 30 14:12:41 crc kubenswrapper[4840]: I0930 14:12:41.059564 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-2ts9b"] Sep 30 14:12:41 crc kubenswrapper[4840]: I0930 14:12:41.066772 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/48123aa4-9fd6-40d9-9d08-f82a623b068a-ovsdbserver-sb\") pod \"48123aa4-9fd6-40d9-9d08-f82a623b068a\" (UID: \"48123aa4-9fd6-40d9-9d08-f82a623b068a\") " Sep 30 14:12:41 crc kubenswrapper[4840]: I0930 14:12:41.066873 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48123aa4-9fd6-40d9-9d08-f82a623b068a-config\") pod \"48123aa4-9fd6-40d9-9d08-f82a623b068a\" (UID: \"48123aa4-9fd6-40d9-9d08-f82a623b068a\") " Sep 30 14:12:41 crc kubenswrapper[4840]: I0930 14:12:41.066922 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48123aa4-9fd6-40d9-9d08-f82a623b068a-dns-svc\") pod \"48123aa4-9fd6-40d9-9d08-f82a623b068a\" (UID: \"48123aa4-9fd6-40d9-9d08-f82a623b068a\") " Sep 30 14:12:41 crc kubenswrapper[4840]: I0930 14:12:41.067022 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/48123aa4-9fd6-40d9-9d08-f82a623b068a-dns-swift-storage-0\") pod \"48123aa4-9fd6-40d9-9d08-f82a623b068a\" (UID: \"48123aa4-9fd6-40d9-9d08-f82a623b068a\") " Sep 30 14:12:41 crc kubenswrapper[4840]: I0930 14:12:41.067042 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/48123aa4-9fd6-40d9-9d08-f82a623b068a-ovsdbserver-nb\") pod \"48123aa4-9fd6-40d9-9d08-f82a623b068a\" (UID: \"48123aa4-9fd6-40d9-9d08-f82a623b068a\") " Sep 30 14:12:41 crc kubenswrapper[4840]: I0930 14:12:41.067091 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbbgz\" (UniqueName: \"kubernetes.io/projected/48123aa4-9fd6-40d9-9d08-f82a623b068a-kube-api-access-bbbgz\") pod \"48123aa4-9fd6-40d9-9d08-f82a623b068a\" (UID: \"48123aa4-9fd6-40d9-9d08-f82a623b068a\") " Sep 30 14:12:41 crc kubenswrapper[4840]: I0930 14:12:41.070380 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 14:12:41 crc kubenswrapper[4840]: I0930 14:12:41.074501 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48123aa4-9fd6-40d9-9d08-f82a623b068a-kube-api-access-bbbgz" (OuterVolumeSpecName: "kube-api-access-bbbgz") pod "48123aa4-9fd6-40d9-9d08-f82a623b068a" (UID: "48123aa4-9fd6-40d9-9d08-f82a623b068a"). InnerVolumeSpecName "kube-api-access-bbbgz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:12:41 crc kubenswrapper[4840]: I0930 14:12:41.103736 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:12:41 crc kubenswrapper[4840]: I0930 14:12:41.109829 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48123aa4-9fd6-40d9-9d08-f82a623b068a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "48123aa4-9fd6-40d9-9d08-f82a623b068a" (UID: "48123aa4-9fd6-40d9-9d08-f82a623b068a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:12:41 crc kubenswrapper[4840]: W0930 14:12:41.114773 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod16a2c602_bcd8_4f85_ab33_9325d5c3e9cc.slice/crio-fb4873ff1285bfd9c0cd64abfd617350b3890343bfc8cbd8bedd62f43a1099ff WatchSource:0}: Error finding container fb4873ff1285bfd9c0cd64abfd617350b3890343bfc8cbd8bedd62f43a1099ff: Status 404 returned error can't find the container with id fb4873ff1285bfd9c0cd64abfd617350b3890343bfc8cbd8bedd62f43a1099ff Sep 30 14:12:41 crc kubenswrapper[4840]: I0930 14:12:41.118974 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48123aa4-9fd6-40d9-9d08-f82a623b068a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "48123aa4-9fd6-40d9-9d08-f82a623b068a" (UID: "48123aa4-9fd6-40d9-9d08-f82a623b068a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:12:41 crc kubenswrapper[4840]: I0930 14:12:41.126063 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48123aa4-9fd6-40d9-9d08-f82a623b068a-config" (OuterVolumeSpecName: "config") pod "48123aa4-9fd6-40d9-9d08-f82a623b068a" (UID: "48123aa4-9fd6-40d9-9d08-f82a623b068a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:12:41 crc kubenswrapper[4840]: I0930 14:12:41.143644 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48123aa4-9fd6-40d9-9d08-f82a623b068a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "48123aa4-9fd6-40d9-9d08-f82a623b068a" (UID: "48123aa4-9fd6-40d9-9d08-f82a623b068a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:12:41 crc kubenswrapper[4840]: I0930 14:12:41.159967 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48123aa4-9fd6-40d9-9d08-f82a623b068a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "48123aa4-9fd6-40d9-9d08-f82a623b068a" (UID: "48123aa4-9fd6-40d9-9d08-f82a623b068a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:12:41 crc kubenswrapper[4840]: I0930 14:12:41.168522 4840 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48123aa4-9fd6-40d9-9d08-f82a623b068a-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:41 crc kubenswrapper[4840]: I0930 14:12:41.168567 4840 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48123aa4-9fd6-40d9-9d08-f82a623b068a-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:41 crc kubenswrapper[4840]: I0930 14:12:41.168578 4840 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/48123aa4-9fd6-40d9-9d08-f82a623b068a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:41 crc kubenswrapper[4840]: I0930 14:12:41.168595 4840 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/48123aa4-9fd6-40d9-9d08-f82a623b068a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:41 crc kubenswrapper[4840]: I0930 14:12:41.168603 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbbgz\" (UniqueName: \"kubernetes.io/projected/48123aa4-9fd6-40d9-9d08-f82a623b068a-kube-api-access-bbbgz\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:41 crc kubenswrapper[4840]: I0930 14:12:41.168611 4840 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/48123aa4-9fd6-40d9-9d08-f82a623b068a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:41 crc kubenswrapper[4840]: I0930 14:12:41.707918 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f785c99fc-fpf4g" event={"ID":"47067ffa-fc01-48b4-abcd-4405053eadca","Type":"ContainerStarted","Data":"6102f664417c2451768375070aac7093a8c91d3a91dbc55122b2021866de56ca"} Sep 30 14:12:41 crc kubenswrapper[4840]: I0930 14:12:41.710610 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-nplct" event={"ID":"85509efb-abc4-4649-a320-7b9de5487180","Type":"ContainerStarted","Data":"f4c754f0714009dfde0d74882248a19e96c17147f4c8a912359797cef63027ca"} Sep 30 14:12:41 crc kubenswrapper[4840]: I0930 14:12:41.712756 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc","Type":"ContainerStarted","Data":"fb4873ff1285bfd9c0cd64abfd617350b3890343bfc8cbd8bedd62f43a1099ff"} Sep 30 14:12:41 crc kubenswrapper[4840]: I0930 14:12:41.717587 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-sp7q5" event={"ID":"ad748b83-4edd-4545-9386-032185c531c5","Type":"ContainerStarted","Data":"82168598b29a794bbf74c3fee4e2676ee2b702525cf44ada0f7498f39206978d"} Sep 30 14:12:41 crc kubenswrapper[4840]: I0930 14:12:41.722692 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8c5d999e-175b-426d-877c-4a698218c363","Type":"ContainerStarted","Data":"853cecaa545c73c00b20e0fbbd636781a1abb5b594f1aa4c466727a9f1373ce0"} Sep 30 14:12:41 crc kubenswrapper[4840]: I0930 14:12:41.728004 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-nplct" podStartSLOduration=2.727988122 podStartE2EDuration="2.727988122s" podCreationTimestamp="2025-09-30 14:12:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:12:41.725801265 +0000 UTC m=+990.354887688" watchObservedRunningTime="2025-09-30 14:12:41.727988122 +0000 UTC m=+990.357074545" Sep 30 14:12:41 crc kubenswrapper[4840]: I0930 14:12:41.734798 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-9vd76" event={"ID":"242d13bb-6113-421b-83d1-fcccf91a9c68","Type":"ContainerStarted","Data":"e7ae2c59a20eb580f9d27eb067fb6d3b0b51bff398919efd4d34d75f936615d9"} Sep 30 14:12:41 crc kubenswrapper[4840]: I0930 14:12:41.740976 4840 generic.go:334] "Generic (PLEG): container finished" podID="b73e88ee-cb25-466d-9860-0dd54a1c1f4c" containerID="67c24bcbe6a8e03ad93ea8400fa9337d59ee9b410dc1bf72342e0e16d1c8942a" exitCode=0 Sep 30 14:12:41 crc kubenswrapper[4840]: I0930 14:12:41.741030 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-gxf58" event={"ID":"b73e88ee-cb25-466d-9860-0dd54a1c1f4c","Type":"ContainerDied","Data":"67c24bcbe6a8e03ad93ea8400fa9337d59ee9b410dc1bf72342e0e16d1c8942a"} Sep 30 14:12:41 crc kubenswrapper[4840]: I0930 14:12:41.741051 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-gxf58" event={"ID":"b73e88ee-cb25-466d-9860-0dd54a1c1f4c","Type":"ContainerStarted","Data":"b2c37df544a67982abf51985e156af05ad950bcc25628ce8bd5c54117c6d35eb"} Sep 30 14:12:41 crc kubenswrapper[4840]: I0930 14:12:41.775783 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-chrgb" event={"ID":"48123aa4-9fd6-40d9-9d08-f82a623b068a","Type":"ContainerDied","Data":"631461d277dde56e8837ec1d908468f4d572fac6cb30373c8e69c807cdbc02ca"} Sep 30 14:12:41 crc kubenswrapper[4840]: I0930 14:12:41.775834 4840 scope.go:117] "RemoveContainer" containerID="7b5e16fbefb4004f06014a8628f8180f0b562df2c673b5dde92dbbadff996f16" Sep 30 14:12:41 crc kubenswrapper[4840]: I0930 14:12:41.775927 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-chrgb" Sep 30 14:12:41 crc kubenswrapper[4840]: I0930 14:12:41.785102 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e4a730c4-623d-4ced-a687-bec2d627eae9","Type":"ContainerStarted","Data":"d36952b559a81c75053f08f503612e3c98121ccd14059d960e29d7496be69ca8"} Sep 30 14:12:42 crc kubenswrapper[4840]: I0930 14:12:42.002692 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-chrgb"] Sep 30 14:12:42 crc kubenswrapper[4840]: I0930 14:12:42.024021 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-chrgb"] Sep 30 14:12:42 crc kubenswrapper[4840]: I0930 14:12:42.039133 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5cc547b489-2p58b"] Sep 30 14:12:42 crc kubenswrapper[4840]: I0930 14:12:42.064637 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:12:42 crc kubenswrapper[4840]: I0930 14:12:42.091010 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-788d88bd6f-l44pw"] Sep 30 14:12:42 crc kubenswrapper[4840]: E0930 14:12:42.091430 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f130d775-9523-40f2-8ccd-24ab0dedb3b6" containerName="dnsmasq-dns" Sep 30 14:12:42 crc kubenswrapper[4840]: I0930 14:12:42.091443 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="f130d775-9523-40f2-8ccd-24ab0dedb3b6" containerName="dnsmasq-dns" Sep 30 14:12:42 crc kubenswrapper[4840]: E0930 14:12:42.091469 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48123aa4-9fd6-40d9-9d08-f82a623b068a" containerName="init" Sep 30 14:12:42 crc kubenswrapper[4840]: I0930 14:12:42.091475 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="48123aa4-9fd6-40d9-9d08-f82a623b068a" containerName="init" Sep 30 14:12:42 crc kubenswrapper[4840]: E0930 14:12:42.091483 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f130d775-9523-40f2-8ccd-24ab0dedb3b6" containerName="init" Sep 30 14:12:42 crc kubenswrapper[4840]: I0930 14:12:42.091489 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="f130d775-9523-40f2-8ccd-24ab0dedb3b6" containerName="init" Sep 30 14:12:42 crc kubenswrapper[4840]: I0930 14:12:42.091696 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="48123aa4-9fd6-40d9-9d08-f82a623b068a" containerName="init" Sep 30 14:12:42 crc kubenswrapper[4840]: I0930 14:12:42.091719 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="f130d775-9523-40f2-8ccd-24ab0dedb3b6" containerName="dnsmasq-dns" Sep 30 14:12:42 crc kubenswrapper[4840]: I0930 14:12:42.092671 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-788d88bd6f-l44pw" Sep 30 14:12:42 crc kubenswrapper[4840]: I0930 14:12:42.109534 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-788d88bd6f-l44pw"] Sep 30 14:12:42 crc kubenswrapper[4840]: I0930 14:12:42.114712 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:12:42 crc kubenswrapper[4840]: I0930 14:12:42.152420 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48123aa4-9fd6-40d9-9d08-f82a623b068a" path="/var/lib/kubelet/pods/48123aa4-9fd6-40d9-9d08-f82a623b068a/volumes" Sep 30 14:12:42 crc kubenswrapper[4840]: I0930 14:12:42.152946 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f130d775-9523-40f2-8ccd-24ab0dedb3b6" path="/var/lib/kubelet/pods/f130d775-9523-40f2-8ccd-24ab0dedb3b6/volumes" Sep 30 14:12:42 crc kubenswrapper[4840]: I0930 14:12:42.153454 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 14:12:42 crc kubenswrapper[4840]: I0930 14:12:42.195780 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e141523f-ce9b-4626-b155-b0a82a3f6dc6-scripts\") pod \"horizon-788d88bd6f-l44pw\" (UID: \"e141523f-ce9b-4626-b155-b0a82a3f6dc6\") " pod="openstack/horizon-788d88bd6f-l44pw" Sep 30 14:12:42 crc kubenswrapper[4840]: I0930 14:12:42.195861 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e141523f-ce9b-4626-b155-b0a82a3f6dc6-horizon-secret-key\") pod \"horizon-788d88bd6f-l44pw\" (UID: \"e141523f-ce9b-4626-b155-b0a82a3f6dc6\") " pod="openstack/horizon-788d88bd6f-l44pw" Sep 30 14:12:42 crc kubenswrapper[4840]: I0930 14:12:42.195888 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e141523f-ce9b-4626-b155-b0a82a3f6dc6-logs\") pod \"horizon-788d88bd6f-l44pw\" (UID: \"e141523f-ce9b-4626-b155-b0a82a3f6dc6\") " pod="openstack/horizon-788d88bd6f-l44pw" Sep 30 14:12:42 crc kubenswrapper[4840]: I0930 14:12:42.195956 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fp459\" (UniqueName: \"kubernetes.io/projected/e141523f-ce9b-4626-b155-b0a82a3f6dc6-kube-api-access-fp459\") pod \"horizon-788d88bd6f-l44pw\" (UID: \"e141523f-ce9b-4626-b155-b0a82a3f6dc6\") " pod="openstack/horizon-788d88bd6f-l44pw" Sep 30 14:12:42 crc kubenswrapper[4840]: I0930 14:12:42.195978 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e141523f-ce9b-4626-b155-b0a82a3f6dc6-config-data\") pod \"horizon-788d88bd6f-l44pw\" (UID: \"e141523f-ce9b-4626-b155-b0a82a3f6dc6\") " pod="openstack/horizon-788d88bd6f-l44pw" Sep 30 14:12:42 crc kubenswrapper[4840]: I0930 14:12:42.301263 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fp459\" (UniqueName: \"kubernetes.io/projected/e141523f-ce9b-4626-b155-b0a82a3f6dc6-kube-api-access-fp459\") pod \"horizon-788d88bd6f-l44pw\" (UID: \"e141523f-ce9b-4626-b155-b0a82a3f6dc6\") " pod="openstack/horizon-788d88bd6f-l44pw" Sep 30 14:12:42 crc kubenswrapper[4840]: I0930 14:12:42.301363 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e141523f-ce9b-4626-b155-b0a82a3f6dc6-config-data\") pod \"horizon-788d88bd6f-l44pw\" (UID: \"e141523f-ce9b-4626-b155-b0a82a3f6dc6\") " pod="openstack/horizon-788d88bd6f-l44pw" Sep 30 14:12:42 crc kubenswrapper[4840]: I0930 14:12:42.301669 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e141523f-ce9b-4626-b155-b0a82a3f6dc6-scripts\") pod \"horizon-788d88bd6f-l44pw\" (UID: \"e141523f-ce9b-4626-b155-b0a82a3f6dc6\") " pod="openstack/horizon-788d88bd6f-l44pw" Sep 30 14:12:42 crc kubenswrapper[4840]: I0930 14:12:42.301808 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e141523f-ce9b-4626-b155-b0a82a3f6dc6-horizon-secret-key\") pod \"horizon-788d88bd6f-l44pw\" (UID: \"e141523f-ce9b-4626-b155-b0a82a3f6dc6\") " pod="openstack/horizon-788d88bd6f-l44pw" Sep 30 14:12:42 crc kubenswrapper[4840]: I0930 14:12:42.301842 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e141523f-ce9b-4626-b155-b0a82a3f6dc6-logs\") pod \"horizon-788d88bd6f-l44pw\" (UID: \"e141523f-ce9b-4626-b155-b0a82a3f6dc6\") " pod="openstack/horizon-788d88bd6f-l44pw" Sep 30 14:12:42 crc kubenswrapper[4840]: I0930 14:12:42.302312 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e141523f-ce9b-4626-b155-b0a82a3f6dc6-logs\") pod \"horizon-788d88bd6f-l44pw\" (UID: \"e141523f-ce9b-4626-b155-b0a82a3f6dc6\") " pod="openstack/horizon-788d88bd6f-l44pw" Sep 30 14:12:42 crc kubenswrapper[4840]: I0930 14:12:42.303092 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e141523f-ce9b-4626-b155-b0a82a3f6dc6-scripts\") pod \"horizon-788d88bd6f-l44pw\" (UID: \"e141523f-ce9b-4626-b155-b0a82a3f6dc6\") " pod="openstack/horizon-788d88bd6f-l44pw" Sep 30 14:12:42 crc kubenswrapper[4840]: I0930 14:12:42.304458 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e141523f-ce9b-4626-b155-b0a82a3f6dc6-config-data\") pod \"horizon-788d88bd6f-l44pw\" (UID: \"e141523f-ce9b-4626-b155-b0a82a3f6dc6\") " pod="openstack/horizon-788d88bd6f-l44pw" Sep 30 14:12:42 crc kubenswrapper[4840]: I0930 14:12:42.311136 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e141523f-ce9b-4626-b155-b0a82a3f6dc6-horizon-secret-key\") pod \"horizon-788d88bd6f-l44pw\" (UID: \"e141523f-ce9b-4626-b155-b0a82a3f6dc6\") " pod="openstack/horizon-788d88bd6f-l44pw" Sep 30 14:12:42 crc kubenswrapper[4840]: I0930 14:12:42.326758 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fp459\" (UniqueName: \"kubernetes.io/projected/e141523f-ce9b-4626-b155-b0a82a3f6dc6-kube-api-access-fp459\") pod \"horizon-788d88bd6f-l44pw\" (UID: \"e141523f-ce9b-4626-b155-b0a82a3f6dc6\") " pod="openstack/horizon-788d88bd6f-l44pw" Sep 30 14:12:42 crc kubenswrapper[4840]: I0930 14:12:42.468594 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-788d88bd6f-l44pw" Sep 30 14:12:42 crc kubenswrapper[4840]: I0930 14:12:42.860805 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc","Type":"ContainerStarted","Data":"a2afef429e08db82a843b5ebc0fa79403f00fe910e2d02f574d9c35fdede612f"} Sep 30 14:12:42 crc kubenswrapper[4840]: I0930 14:12:42.864226 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8c5d999e-175b-426d-877c-4a698218c363","Type":"ContainerStarted","Data":"7aa98fc6015e602ca6609d69a712c47542cdb3a0afd32ad496cfb5db2098b7a7"} Sep 30 14:12:42 crc kubenswrapper[4840]: I0930 14:12:42.868835 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-gxf58" event={"ID":"b73e88ee-cb25-466d-9860-0dd54a1c1f4c","Type":"ContainerStarted","Data":"8f1dc45f77092a5778340c020c0952b71db3f7fad8ff85b132ad5c8cc6e7a5c5"} Sep 30 14:12:42 crc kubenswrapper[4840]: I0930 14:12:42.869273 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8b5c85b87-gxf58" Sep 30 14:12:42 crc kubenswrapper[4840]: I0930 14:12:42.895231 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8b5c85b87-gxf58" podStartSLOduration=3.895216581 podStartE2EDuration="3.895216581s" podCreationTimestamp="2025-09-30 14:12:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:12:42.894002079 +0000 UTC m=+991.523088522" watchObservedRunningTime="2025-09-30 14:12:42.895216581 +0000 UTC m=+991.524303004" Sep 30 14:12:43 crc kubenswrapper[4840]: I0930 14:12:43.056193 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-9hcwn"] Sep 30 14:12:43 crc kubenswrapper[4840]: I0930 14:12:43.057320 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9hcwn" Sep 30 14:12:43 crc kubenswrapper[4840]: I0930 14:12:43.060224 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-g5svs" Sep 30 14:12:43 crc kubenswrapper[4840]: I0930 14:12:43.060451 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Sep 30 14:12:43 crc kubenswrapper[4840]: I0930 14:12:43.087926 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-9hcwn"] Sep 30 14:12:43 crc kubenswrapper[4840]: I0930 14:12:43.127975 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-788d88bd6f-l44pw"] Sep 30 14:12:43 crc kubenswrapper[4840]: W0930 14:12:43.144754 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode141523f_ce9b_4626_b155_b0a82a3f6dc6.slice/crio-99172bb3072324501f5d649ac3715c1f111ca64410df87296a892bccd91a51b5 WatchSource:0}: Error finding container 99172bb3072324501f5d649ac3715c1f111ca64410df87296a892bccd91a51b5: Status 404 returned error can't find the container with id 99172bb3072324501f5d649ac3715c1f111ca64410df87296a892bccd91a51b5 Sep 30 14:12:43 crc kubenswrapper[4840]: I0930 14:12:43.217040 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c465939-fcbe-44b2-b7ae-a24b5a581dbc-combined-ca-bundle\") pod \"barbican-db-sync-9hcwn\" (UID: \"2c465939-fcbe-44b2-b7ae-a24b5a581dbc\") " pod="openstack/barbican-db-sync-9hcwn" Sep 30 14:12:43 crc kubenswrapper[4840]: I0930 14:12:43.217120 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4l6jg\" (UniqueName: \"kubernetes.io/projected/2c465939-fcbe-44b2-b7ae-a24b5a581dbc-kube-api-access-4l6jg\") pod \"barbican-db-sync-9hcwn\" (UID: \"2c465939-fcbe-44b2-b7ae-a24b5a581dbc\") " pod="openstack/barbican-db-sync-9hcwn" Sep 30 14:12:43 crc kubenswrapper[4840]: I0930 14:12:43.217208 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2c465939-fcbe-44b2-b7ae-a24b5a581dbc-db-sync-config-data\") pod \"barbican-db-sync-9hcwn\" (UID: \"2c465939-fcbe-44b2-b7ae-a24b5a581dbc\") " pod="openstack/barbican-db-sync-9hcwn" Sep 30 14:12:43 crc kubenswrapper[4840]: I0930 14:12:43.318728 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4l6jg\" (UniqueName: \"kubernetes.io/projected/2c465939-fcbe-44b2-b7ae-a24b5a581dbc-kube-api-access-4l6jg\") pod \"barbican-db-sync-9hcwn\" (UID: \"2c465939-fcbe-44b2-b7ae-a24b5a581dbc\") " pod="openstack/barbican-db-sync-9hcwn" Sep 30 14:12:43 crc kubenswrapper[4840]: I0930 14:12:43.318823 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2c465939-fcbe-44b2-b7ae-a24b5a581dbc-db-sync-config-data\") pod \"barbican-db-sync-9hcwn\" (UID: \"2c465939-fcbe-44b2-b7ae-a24b5a581dbc\") " pod="openstack/barbican-db-sync-9hcwn" Sep 30 14:12:43 crc kubenswrapper[4840]: I0930 14:12:43.318953 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c465939-fcbe-44b2-b7ae-a24b5a581dbc-combined-ca-bundle\") pod \"barbican-db-sync-9hcwn\" (UID: \"2c465939-fcbe-44b2-b7ae-a24b5a581dbc\") " pod="openstack/barbican-db-sync-9hcwn" Sep 30 14:12:43 crc kubenswrapper[4840]: I0930 14:12:43.331883 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2c465939-fcbe-44b2-b7ae-a24b5a581dbc-db-sync-config-data\") pod \"barbican-db-sync-9hcwn\" (UID: \"2c465939-fcbe-44b2-b7ae-a24b5a581dbc\") " pod="openstack/barbican-db-sync-9hcwn" Sep 30 14:12:43 crc kubenswrapper[4840]: I0930 14:12:43.336516 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c465939-fcbe-44b2-b7ae-a24b5a581dbc-combined-ca-bundle\") pod \"barbican-db-sync-9hcwn\" (UID: \"2c465939-fcbe-44b2-b7ae-a24b5a581dbc\") " pod="openstack/barbican-db-sync-9hcwn" Sep 30 14:12:43 crc kubenswrapper[4840]: I0930 14:12:43.342083 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4l6jg\" (UniqueName: \"kubernetes.io/projected/2c465939-fcbe-44b2-b7ae-a24b5a581dbc-kube-api-access-4l6jg\") pod \"barbican-db-sync-9hcwn\" (UID: \"2c465939-fcbe-44b2-b7ae-a24b5a581dbc\") " pod="openstack/barbican-db-sync-9hcwn" Sep 30 14:12:43 crc kubenswrapper[4840]: I0930 14:12:43.395271 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9hcwn" Sep 30 14:12:43 crc kubenswrapper[4840]: I0930 14:12:43.684084 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-9hcwn"] Sep 30 14:12:43 crc kubenswrapper[4840]: I0930 14:12:43.917411 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc","Type":"ContainerStarted","Data":"6898fce6551a8735dee36a89b338612548caef2e3441fd6383b3783ca715ddfc"} Sep 30 14:12:43 crc kubenswrapper[4840]: I0930 14:12:43.917843 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="16a2c602-bcd8-4f85-ab33-9325d5c3e9cc" containerName="glance-log" containerID="cri-o://a2afef429e08db82a843b5ebc0fa79403f00fe910e2d02f574d9c35fdede612f" gracePeriod=30 Sep 30 14:12:43 crc kubenswrapper[4840]: I0930 14:12:43.918315 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="16a2c602-bcd8-4f85-ab33-9325d5c3e9cc" containerName="glance-httpd" containerID="cri-o://6898fce6551a8735dee36a89b338612548caef2e3441fd6383b3783ca715ddfc" gracePeriod=30 Sep 30 14:12:43 crc kubenswrapper[4840]: I0930 14:12:43.928027 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8c5d999e-175b-426d-877c-4a698218c363","Type":"ContainerStarted","Data":"70029353c5195b1b6f43700b02ee9bc10d4d6aaf1ea4ac7ca79015b811c337eb"} Sep 30 14:12:43 crc kubenswrapper[4840]: I0930 14:12:43.928153 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="8c5d999e-175b-426d-877c-4a698218c363" containerName="glance-log" containerID="cri-o://7aa98fc6015e602ca6609d69a712c47542cdb3a0afd32ad496cfb5db2098b7a7" gracePeriod=30 Sep 30 14:12:43 crc kubenswrapper[4840]: I0930 14:12:43.928234 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="8c5d999e-175b-426d-877c-4a698218c363" containerName="glance-httpd" containerID="cri-o://70029353c5195b1b6f43700b02ee9bc10d4d6aaf1ea4ac7ca79015b811c337eb" gracePeriod=30 Sep 30 14:12:43 crc kubenswrapper[4840]: I0930 14:12:43.951185 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.951168108 podStartE2EDuration="4.951168108s" podCreationTimestamp="2025-09-30 14:12:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:12:43.948440157 +0000 UTC m=+992.577526580" watchObservedRunningTime="2025-09-30 14:12:43.951168108 +0000 UTC m=+992.580254531" Sep 30 14:12:43 crc kubenswrapper[4840]: I0930 14:12:43.969512 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-788d88bd6f-l44pw" event={"ID":"e141523f-ce9b-4626-b155-b0a82a3f6dc6","Type":"ContainerStarted","Data":"99172bb3072324501f5d649ac3715c1f111ca64410df87296a892bccd91a51b5"} Sep 30 14:12:43 crc kubenswrapper[4840]: I0930 14:12:43.972317 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9hcwn" event={"ID":"2c465939-fcbe-44b2-b7ae-a24b5a581dbc","Type":"ContainerStarted","Data":"3958b3f15ef2941ffd7ae2a69bdb6a938691349e2395edb10287002e93dd7857"} Sep 30 14:12:43 crc kubenswrapper[4840]: I0930 14:12:43.994045 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.994027826 podStartE2EDuration="4.994027826s" podCreationTimestamp="2025-09-30 14:12:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:12:43.99110725 +0000 UTC m=+992.620193673" watchObservedRunningTime="2025-09-30 14:12:43.994027826 +0000 UTC m=+992.623114249" Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.594431 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.754424 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8c5d999e-175b-426d-877c-4a698218c363-httpd-run\") pod \"8c5d999e-175b-426d-877c-4a698218c363\" (UID: \"8c5d999e-175b-426d-877c-4a698218c363\") " Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.754833 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c5d999e-175b-426d-877c-4a698218c363-combined-ca-bundle\") pod \"8c5d999e-175b-426d-877c-4a698218c363\" (UID: \"8c5d999e-175b-426d-877c-4a698218c363\") " Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.754903 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c5d999e-175b-426d-877c-4a698218c363-public-tls-certs\") pod \"8c5d999e-175b-426d-877c-4a698218c363\" (UID: \"8c5d999e-175b-426d-877c-4a698218c363\") " Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.754934 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c5d999e-175b-426d-877c-4a698218c363-config-data\") pod \"8c5d999e-175b-426d-877c-4a698218c363\" (UID: \"8c5d999e-175b-426d-877c-4a698218c363\") " Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.755003 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c5d999e-175b-426d-877c-4a698218c363-logs\") pod \"8c5d999e-175b-426d-877c-4a698218c363\" (UID: \"8c5d999e-175b-426d-877c-4a698218c363\") " Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.755108 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pskrm\" (UniqueName: \"kubernetes.io/projected/8c5d999e-175b-426d-877c-4a698218c363-kube-api-access-pskrm\") pod \"8c5d999e-175b-426d-877c-4a698218c363\" (UID: \"8c5d999e-175b-426d-877c-4a698218c363\") " Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.755184 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"8c5d999e-175b-426d-877c-4a698218c363\" (UID: \"8c5d999e-175b-426d-877c-4a698218c363\") " Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.755215 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c5d999e-175b-426d-877c-4a698218c363-scripts\") pod \"8c5d999e-175b-426d-877c-4a698218c363\" (UID: \"8c5d999e-175b-426d-877c-4a698218c363\") " Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.756827 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c5d999e-175b-426d-877c-4a698218c363-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8c5d999e-175b-426d-877c-4a698218c363" (UID: "8c5d999e-175b-426d-877c-4a698218c363"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.757231 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c5d999e-175b-426d-877c-4a698218c363-logs" (OuterVolumeSpecName: "logs") pod "8c5d999e-175b-426d-877c-4a698218c363" (UID: "8c5d999e-175b-426d-877c-4a698218c363"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.765727 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.770830 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c5d999e-175b-426d-877c-4a698218c363-kube-api-access-pskrm" (OuterVolumeSpecName: "kube-api-access-pskrm") pod "8c5d999e-175b-426d-877c-4a698218c363" (UID: "8c5d999e-175b-426d-877c-4a698218c363"). InnerVolumeSpecName "kube-api-access-pskrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.777609 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "8c5d999e-175b-426d-877c-4a698218c363" (UID: "8c5d999e-175b-426d-877c-4a698218c363"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.777920 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c5d999e-175b-426d-877c-4a698218c363-scripts" (OuterVolumeSpecName: "scripts") pod "8c5d999e-175b-426d-877c-4a698218c363" (UID: "8c5d999e-175b-426d-877c-4a698218c363"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.819012 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c5d999e-175b-426d-877c-4a698218c363-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8c5d999e-175b-426d-877c-4a698218c363" (UID: "8c5d999e-175b-426d-877c-4a698218c363"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.862688 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c5d999e-175b-426d-877c-4a698218c363-config-data" (OuterVolumeSpecName: "config-data") pod "8c5d999e-175b-426d-877c-4a698218c363" (UID: "8c5d999e-175b-426d-877c-4a698218c363"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.862829 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-internal-tls-certs\") pod \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\" (UID: \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\") " Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.862866 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-combined-ca-bundle\") pod \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\" (UID: \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\") " Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.862899 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-logs\") pod \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\" (UID: \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\") " Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.862923 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-httpd-run\") pod \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\" (UID: \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\") " Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.863052 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\" (UID: \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\") " Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.863111 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-scripts\") pod \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\" (UID: \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\") " Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.863148 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4bfb\" (UniqueName: \"kubernetes.io/projected/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-kube-api-access-f4bfb\") pod \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\" (UID: \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\") " Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.863209 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-config-data\") pod \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\" (UID: \"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc\") " Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.863257 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c5d999e-175b-426d-877c-4a698218c363-config-data\") pod \"8c5d999e-175b-426d-877c-4a698218c363\" (UID: \"8c5d999e-175b-426d-877c-4a698218c363\") " Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.863877 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pskrm\" (UniqueName: \"kubernetes.io/projected/8c5d999e-175b-426d-877c-4a698218c363-kube-api-access-pskrm\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.863905 4840 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.863917 4840 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c5d999e-175b-426d-877c-4a698218c363-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.863927 4840 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8c5d999e-175b-426d-877c-4a698218c363-httpd-run\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.863935 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c5d999e-175b-426d-877c-4a698218c363-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.863965 4840 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c5d999e-175b-426d-877c-4a698218c363-logs\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:44 crc kubenswrapper[4840]: W0930 14:12:44.879125 4840 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/8c5d999e-175b-426d-877c-4a698218c363/volumes/kubernetes.io~secret/config-data Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.879153 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c5d999e-175b-426d-877c-4a698218c363-config-data" (OuterVolumeSpecName: "config-data") pod "8c5d999e-175b-426d-877c-4a698218c363" (UID: "8c5d999e-175b-426d-877c-4a698218c363"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.879547 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-logs" (OuterVolumeSpecName: "logs") pod "16a2c602-bcd8-4f85-ab33-9325d5c3e9cc" (UID: "16a2c602-bcd8-4f85-ab33-9325d5c3e9cc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.879648 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "16a2c602-bcd8-4f85-ab33-9325d5c3e9cc" (UID: "16a2c602-bcd8-4f85-ab33-9325d5c3e9cc"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.880216 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c5d999e-175b-426d-877c-4a698218c363-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "8c5d999e-175b-426d-877c-4a698218c363" (UID: "8c5d999e-175b-426d-877c-4a698218c363"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.890723 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "16a2c602-bcd8-4f85-ab33-9325d5c3e9cc" (UID: "16a2c602-bcd8-4f85-ab33-9325d5c3e9cc"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.891767 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-kube-api-access-f4bfb" (OuterVolumeSpecName: "kube-api-access-f4bfb") pod "16a2c602-bcd8-4f85-ab33-9325d5c3e9cc" (UID: "16a2c602-bcd8-4f85-ab33-9325d5c3e9cc"). InnerVolumeSpecName "kube-api-access-f4bfb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.904707 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-scripts" (OuterVolumeSpecName: "scripts") pod "16a2c602-bcd8-4f85-ab33-9325d5c3e9cc" (UID: "16a2c602-bcd8-4f85-ab33-9325d5c3e9cc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.919589 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "16a2c602-bcd8-4f85-ab33-9325d5c3e9cc" (UID: "16a2c602-bcd8-4f85-ab33-9325d5c3e9cc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.922120 4840 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.939949 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-config-data" (OuterVolumeSpecName: "config-data") pod "16a2c602-bcd8-4f85-ab33-9325d5c3e9cc" (UID: "16a2c602-bcd8-4f85-ab33-9325d5c3e9cc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.940256 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "16a2c602-bcd8-4f85-ab33-9325d5c3e9cc" (UID: "16a2c602-bcd8-4f85-ab33-9325d5c3e9cc"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.966233 4840 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.966266 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.966275 4840 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-logs\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.966284 4840 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-httpd-run\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.966315 4840 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.966326 4840 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.966334 4840 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.966345 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4bfb\" (UniqueName: \"kubernetes.io/projected/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-kube-api-access-f4bfb\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.966359 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.966370 4840 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c5d999e-175b-426d-877c-4a698218c363-public-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.966380 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c5d999e-175b-426d-877c-4a698218c363-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.985201 4840 generic.go:334] "Generic (PLEG): container finished" podID="16a2c602-bcd8-4f85-ab33-9325d5c3e9cc" containerID="6898fce6551a8735dee36a89b338612548caef2e3441fd6383b3783ca715ddfc" exitCode=0 Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.985237 4840 generic.go:334] "Generic (PLEG): container finished" podID="16a2c602-bcd8-4f85-ab33-9325d5c3e9cc" containerID="a2afef429e08db82a843b5ebc0fa79403f00fe910e2d02f574d9c35fdede612f" exitCode=143 Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.985295 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc","Type":"ContainerDied","Data":"6898fce6551a8735dee36a89b338612548caef2e3441fd6383b3783ca715ddfc"} Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.985317 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.985328 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc","Type":"ContainerDied","Data":"a2afef429e08db82a843b5ebc0fa79403f00fe910e2d02f574d9c35fdede612f"} Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.985342 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"16a2c602-bcd8-4f85-ab33-9325d5c3e9cc","Type":"ContainerDied","Data":"fb4873ff1285bfd9c0cd64abfd617350b3890343bfc8cbd8bedd62f43a1099ff"} Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.985357 4840 scope.go:117] "RemoveContainer" containerID="6898fce6551a8735dee36a89b338612548caef2e3441fd6383b3783ca715ddfc" Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.989951 4840 generic.go:334] "Generic (PLEG): container finished" podID="8c5d999e-175b-426d-877c-4a698218c363" containerID="70029353c5195b1b6f43700b02ee9bc10d4d6aaf1ea4ac7ca79015b811c337eb" exitCode=143 Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.989996 4840 generic.go:334] "Generic (PLEG): container finished" podID="8c5d999e-175b-426d-877c-4a698218c363" containerID="7aa98fc6015e602ca6609d69a712c47542cdb3a0afd32ad496cfb5db2098b7a7" exitCode=143 Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.990018 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8c5d999e-175b-426d-877c-4a698218c363","Type":"ContainerDied","Data":"70029353c5195b1b6f43700b02ee9bc10d4d6aaf1ea4ac7ca79015b811c337eb"} Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.990048 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8c5d999e-175b-426d-877c-4a698218c363","Type":"ContainerDied","Data":"7aa98fc6015e602ca6609d69a712c47542cdb3a0afd32ad496cfb5db2098b7a7"} Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.990060 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8c5d999e-175b-426d-877c-4a698218c363","Type":"ContainerDied","Data":"853cecaa545c73c00b20e0fbbd636781a1abb5b594f1aa4c466727a9f1373ce0"} Sep 30 14:12:44 crc kubenswrapper[4840]: I0930 14:12:44.990122 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.002010 4840 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.054795 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.062881 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.072526 4840 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.093469 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.101763 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.109689 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:12:45 crc kubenswrapper[4840]: E0930 14:12:45.110117 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16a2c602-bcd8-4f85-ab33-9325d5c3e9cc" containerName="glance-httpd" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.110129 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="16a2c602-bcd8-4f85-ab33-9325d5c3e9cc" containerName="glance-httpd" Sep 30 14:12:45 crc kubenswrapper[4840]: E0930 14:12:45.110139 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16a2c602-bcd8-4f85-ab33-9325d5c3e9cc" containerName="glance-log" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.110146 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="16a2c602-bcd8-4f85-ab33-9325d5c3e9cc" containerName="glance-log" Sep 30 14:12:45 crc kubenswrapper[4840]: E0930 14:12:45.110158 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c5d999e-175b-426d-877c-4a698218c363" containerName="glance-httpd" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.110164 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c5d999e-175b-426d-877c-4a698218c363" containerName="glance-httpd" Sep 30 14:12:45 crc kubenswrapper[4840]: E0930 14:12:45.110178 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c5d999e-175b-426d-877c-4a698218c363" containerName="glance-log" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.110184 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c5d999e-175b-426d-877c-4a698218c363" containerName="glance-log" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.110498 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="16a2c602-bcd8-4f85-ab33-9325d5c3e9cc" containerName="glance-httpd" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.110526 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c5d999e-175b-426d-877c-4a698218c363" containerName="glance-log" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.110540 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="16a2c602-bcd8-4f85-ab33-9325d5c3e9cc" containerName="glance-log" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.110564 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c5d999e-175b-426d-877c-4a698218c363" containerName="glance-httpd" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.112642 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.115219 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.115837 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-7s8kx" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.116723 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.119177 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.141375 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.147964 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.155460 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.156136 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.156241 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.168210 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.279474 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7sqqs\" (UniqueName: \"kubernetes.io/projected/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-kube-api-access-7sqqs\") pod \"glance-default-internal-api-0\" (UID: \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.279637 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.280404 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-logs\") pod \"glance-default-internal-api-0\" (UID: \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.280443 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.280474 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-scripts\") pod \"glance-default-external-api-0\" (UID: \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.280497 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.280673 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.280727 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.280754 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.280786 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.280816 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.280900 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-logs\") pod \"glance-default-external-api-0\" (UID: \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.280924 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.280948 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-config-data\") pod \"glance-default-external-api-0\" (UID: \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.280988 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.281019 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wb5df\" (UniqueName: \"kubernetes.io/projected/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-kube-api-access-wb5df\") pod \"glance-default-external-api-0\" (UID: \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.382900 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.382964 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-logs\") pod \"glance-default-external-api-0\" (UID: \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.383000 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.383023 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-config-data\") pod \"glance-default-external-api-0\" (UID: \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.383049 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.383081 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wb5df\" (UniqueName: \"kubernetes.io/projected/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-kube-api-access-wb5df\") pod \"glance-default-external-api-0\" (UID: \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.383107 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7sqqs\" (UniqueName: \"kubernetes.io/projected/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-kube-api-access-7sqqs\") pod \"glance-default-internal-api-0\" (UID: \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.383167 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.383216 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-logs\") pod \"glance-default-internal-api-0\" (UID: \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.383242 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.383266 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-scripts\") pod \"glance-default-external-api-0\" (UID: \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.383289 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.383355 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.383376 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.383398 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.383425 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.383692 4840 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.385696 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-logs\") pod \"glance-default-external-api-0\" (UID: \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.391239 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.391754 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.391809 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-logs\") pod \"glance-default-internal-api-0\" (UID: \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.391905 4840 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-internal-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.392016 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.392401 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.392602 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.393007 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.398822 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.409542 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-scripts\") pod \"glance-default-external-api-0\" (UID: \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.413079 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7sqqs\" (UniqueName: \"kubernetes.io/projected/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-kube-api-access-7sqqs\") pod \"glance-default-internal-api-0\" (UID: \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.413565 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-config-data\") pod \"glance-default-external-api-0\" (UID: \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.415738 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wb5df\" (UniqueName: \"kubernetes.io/projected/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-kube-api-access-wb5df\") pod \"glance-default-external-api-0\" (UID: \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.419009 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.454130 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.458627 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\") " pod="openstack/glance-default-external-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.491615 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.502903 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.999000 4840 generic.go:334] "Generic (PLEG): container finished" podID="1407ce03-21cb-4f1a-9719-6a9aa88a8973" containerID="9f12e56d9a170038e99c98ffb6f13331191d2fdd43eecac910f50328d12487f7" exitCode=0 Sep 30 14:12:45 crc kubenswrapper[4840]: I0930 14:12:45.999078 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-pxgwt" event={"ID":"1407ce03-21cb-4f1a-9719-6a9aa88a8973","Type":"ContainerDied","Data":"9f12e56d9a170038e99c98ffb6f13331191d2fdd43eecac910f50328d12487f7"} Sep 30 14:12:46 crc kubenswrapper[4840]: I0930 14:12:46.127048 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16a2c602-bcd8-4f85-ab33-9325d5c3e9cc" path="/var/lib/kubelet/pods/16a2c602-bcd8-4f85-ab33-9325d5c3e9cc/volumes" Sep 30 14:12:46 crc kubenswrapper[4840]: I0930 14:12:46.127926 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c5d999e-175b-426d-877c-4a698218c363" path="/var/lib/kubelet/pods/8c5d999e-175b-426d-877c-4a698218c363/volumes" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.226658 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7f785c99fc-fpf4g"] Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.260617 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5595d79c68-dgsth"] Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.264444 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5595d79c68-dgsth" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.273971 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.278092 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.291027 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5595d79c68-dgsth"] Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.331360 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-788d88bd6f-l44pw"] Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.332127 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmn5v\" (UniqueName: \"kubernetes.io/projected/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-kube-api-access-xmn5v\") pod \"horizon-5595d79c68-dgsth\" (UID: \"5e9b87b2-4836-4857-8514-b2dcfe4de1f5\") " pod="openstack/horizon-5595d79c68-dgsth" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.332200 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-horizon-secret-key\") pod \"horizon-5595d79c68-dgsth\" (UID: \"5e9b87b2-4836-4857-8514-b2dcfe4de1f5\") " pod="openstack/horizon-5595d79c68-dgsth" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.332223 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-config-data\") pod \"horizon-5595d79c68-dgsth\" (UID: \"5e9b87b2-4836-4857-8514-b2dcfe4de1f5\") " pod="openstack/horizon-5595d79c68-dgsth" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.332262 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-horizon-tls-certs\") pod \"horizon-5595d79c68-dgsth\" (UID: \"5e9b87b2-4836-4857-8514-b2dcfe4de1f5\") " pod="openstack/horizon-5595d79c68-dgsth" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.332545 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-combined-ca-bundle\") pod \"horizon-5595d79c68-dgsth\" (UID: \"5e9b87b2-4836-4857-8514-b2dcfe4de1f5\") " pod="openstack/horizon-5595d79c68-dgsth" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.332708 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-scripts\") pod \"horizon-5595d79c68-dgsth\" (UID: \"5e9b87b2-4836-4857-8514-b2dcfe4de1f5\") " pod="openstack/horizon-5595d79c68-dgsth" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.332784 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-logs\") pod \"horizon-5595d79c68-dgsth\" (UID: \"5e9b87b2-4836-4857-8514-b2dcfe4de1f5\") " pod="openstack/horizon-5595d79c68-dgsth" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.364381 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.381514 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-94dfcc48d-mlrjd"] Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.390593 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-94dfcc48d-mlrjd" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.391217 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-94dfcc48d-mlrjd"] Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.434289 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-logs\") pod \"horizon-5595d79c68-dgsth\" (UID: \"5e9b87b2-4836-4857-8514-b2dcfe4de1f5\") " pod="openstack/horizon-5595d79c68-dgsth" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.434349 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmn5v\" (UniqueName: \"kubernetes.io/projected/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-kube-api-access-xmn5v\") pod \"horizon-5595d79c68-dgsth\" (UID: \"5e9b87b2-4836-4857-8514-b2dcfe4de1f5\") " pod="openstack/horizon-5595d79c68-dgsth" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.434399 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-horizon-secret-key\") pod \"horizon-5595d79c68-dgsth\" (UID: \"5e9b87b2-4836-4857-8514-b2dcfe4de1f5\") " pod="openstack/horizon-5595d79c68-dgsth" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.434414 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-config-data\") pod \"horizon-5595d79c68-dgsth\" (UID: \"5e9b87b2-4836-4857-8514-b2dcfe4de1f5\") " pod="openstack/horizon-5595d79c68-dgsth" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.434454 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-horizon-tls-certs\") pod \"horizon-5595d79c68-dgsth\" (UID: \"5e9b87b2-4836-4857-8514-b2dcfe4de1f5\") " pod="openstack/horizon-5595d79c68-dgsth" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.434506 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-combined-ca-bundle\") pod \"horizon-5595d79c68-dgsth\" (UID: \"5e9b87b2-4836-4857-8514-b2dcfe4de1f5\") " pod="openstack/horizon-5595d79c68-dgsth" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.434550 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-scripts\") pod \"horizon-5595d79c68-dgsth\" (UID: \"5e9b87b2-4836-4857-8514-b2dcfe4de1f5\") " pod="openstack/horizon-5595d79c68-dgsth" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.435789 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-scripts\") pod \"horizon-5595d79c68-dgsth\" (UID: \"5e9b87b2-4836-4857-8514-b2dcfe4de1f5\") " pod="openstack/horizon-5595d79c68-dgsth" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.436675 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-config-data\") pod \"horizon-5595d79c68-dgsth\" (UID: \"5e9b87b2-4836-4857-8514-b2dcfe4de1f5\") " pod="openstack/horizon-5595d79c68-dgsth" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.436930 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-logs\") pod \"horizon-5595d79c68-dgsth\" (UID: \"5e9b87b2-4836-4857-8514-b2dcfe4de1f5\") " pod="openstack/horizon-5595d79c68-dgsth" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.444563 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-horizon-tls-certs\") pod \"horizon-5595d79c68-dgsth\" (UID: \"5e9b87b2-4836-4857-8514-b2dcfe4de1f5\") " pod="openstack/horizon-5595d79c68-dgsth" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.444864 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-horizon-secret-key\") pod \"horizon-5595d79c68-dgsth\" (UID: \"5e9b87b2-4836-4857-8514-b2dcfe4de1f5\") " pod="openstack/horizon-5595d79c68-dgsth" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.446995 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-combined-ca-bundle\") pod \"horizon-5595d79c68-dgsth\" (UID: \"5e9b87b2-4836-4857-8514-b2dcfe4de1f5\") " pod="openstack/horizon-5595d79c68-dgsth" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.452951 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmn5v\" (UniqueName: \"kubernetes.io/projected/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-kube-api-access-xmn5v\") pod \"horizon-5595d79c68-dgsth\" (UID: \"5e9b87b2-4836-4857-8514-b2dcfe4de1f5\") " pod="openstack/horizon-5595d79c68-dgsth" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.536059 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1138b51b-edf9-42d2-8ea4-4d4b9122eae4-scripts\") pod \"horizon-94dfcc48d-mlrjd\" (UID: \"1138b51b-edf9-42d2-8ea4-4d4b9122eae4\") " pod="openstack/horizon-94dfcc48d-mlrjd" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.536144 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1138b51b-edf9-42d2-8ea4-4d4b9122eae4-combined-ca-bundle\") pod \"horizon-94dfcc48d-mlrjd\" (UID: \"1138b51b-edf9-42d2-8ea4-4d4b9122eae4\") " pod="openstack/horizon-94dfcc48d-mlrjd" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.536168 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wb2sf\" (UniqueName: \"kubernetes.io/projected/1138b51b-edf9-42d2-8ea4-4d4b9122eae4-kube-api-access-wb2sf\") pod \"horizon-94dfcc48d-mlrjd\" (UID: \"1138b51b-edf9-42d2-8ea4-4d4b9122eae4\") " pod="openstack/horizon-94dfcc48d-mlrjd" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.536206 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1138b51b-edf9-42d2-8ea4-4d4b9122eae4-config-data\") pod \"horizon-94dfcc48d-mlrjd\" (UID: \"1138b51b-edf9-42d2-8ea4-4d4b9122eae4\") " pod="openstack/horizon-94dfcc48d-mlrjd" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.536264 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/1138b51b-edf9-42d2-8ea4-4d4b9122eae4-horizon-tls-certs\") pod \"horizon-94dfcc48d-mlrjd\" (UID: \"1138b51b-edf9-42d2-8ea4-4d4b9122eae4\") " pod="openstack/horizon-94dfcc48d-mlrjd" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.536371 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1138b51b-edf9-42d2-8ea4-4d4b9122eae4-logs\") pod \"horizon-94dfcc48d-mlrjd\" (UID: \"1138b51b-edf9-42d2-8ea4-4d4b9122eae4\") " pod="openstack/horizon-94dfcc48d-mlrjd" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.536450 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1138b51b-edf9-42d2-8ea4-4d4b9122eae4-horizon-secret-key\") pod \"horizon-94dfcc48d-mlrjd\" (UID: \"1138b51b-edf9-42d2-8ea4-4d4b9122eae4\") " pod="openstack/horizon-94dfcc48d-mlrjd" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.593443 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5595d79c68-dgsth" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.638493 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1138b51b-edf9-42d2-8ea4-4d4b9122eae4-scripts\") pod \"horizon-94dfcc48d-mlrjd\" (UID: \"1138b51b-edf9-42d2-8ea4-4d4b9122eae4\") " pod="openstack/horizon-94dfcc48d-mlrjd" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.638594 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1138b51b-edf9-42d2-8ea4-4d4b9122eae4-combined-ca-bundle\") pod \"horizon-94dfcc48d-mlrjd\" (UID: \"1138b51b-edf9-42d2-8ea4-4d4b9122eae4\") " pod="openstack/horizon-94dfcc48d-mlrjd" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.638651 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wb2sf\" (UniqueName: \"kubernetes.io/projected/1138b51b-edf9-42d2-8ea4-4d4b9122eae4-kube-api-access-wb2sf\") pod \"horizon-94dfcc48d-mlrjd\" (UID: \"1138b51b-edf9-42d2-8ea4-4d4b9122eae4\") " pod="openstack/horizon-94dfcc48d-mlrjd" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.639101 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1138b51b-edf9-42d2-8ea4-4d4b9122eae4-config-data\") pod \"horizon-94dfcc48d-mlrjd\" (UID: \"1138b51b-edf9-42d2-8ea4-4d4b9122eae4\") " pod="openstack/horizon-94dfcc48d-mlrjd" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.639629 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/1138b51b-edf9-42d2-8ea4-4d4b9122eae4-horizon-tls-certs\") pod \"horizon-94dfcc48d-mlrjd\" (UID: \"1138b51b-edf9-42d2-8ea4-4d4b9122eae4\") " pod="openstack/horizon-94dfcc48d-mlrjd" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.639896 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1138b51b-edf9-42d2-8ea4-4d4b9122eae4-logs\") pod \"horizon-94dfcc48d-mlrjd\" (UID: \"1138b51b-edf9-42d2-8ea4-4d4b9122eae4\") " pod="openstack/horizon-94dfcc48d-mlrjd" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.639984 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1138b51b-edf9-42d2-8ea4-4d4b9122eae4-horizon-secret-key\") pod \"horizon-94dfcc48d-mlrjd\" (UID: \"1138b51b-edf9-42d2-8ea4-4d4b9122eae4\") " pod="openstack/horizon-94dfcc48d-mlrjd" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.640228 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1138b51b-edf9-42d2-8ea4-4d4b9122eae4-scripts\") pod \"horizon-94dfcc48d-mlrjd\" (UID: \"1138b51b-edf9-42d2-8ea4-4d4b9122eae4\") " pod="openstack/horizon-94dfcc48d-mlrjd" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.640294 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1138b51b-edf9-42d2-8ea4-4d4b9122eae4-logs\") pod \"horizon-94dfcc48d-mlrjd\" (UID: \"1138b51b-edf9-42d2-8ea4-4d4b9122eae4\") " pod="openstack/horizon-94dfcc48d-mlrjd" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.640726 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1138b51b-edf9-42d2-8ea4-4d4b9122eae4-config-data\") pod \"horizon-94dfcc48d-mlrjd\" (UID: \"1138b51b-edf9-42d2-8ea4-4d4b9122eae4\") " pod="openstack/horizon-94dfcc48d-mlrjd" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.643878 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/1138b51b-edf9-42d2-8ea4-4d4b9122eae4-horizon-tls-certs\") pod \"horizon-94dfcc48d-mlrjd\" (UID: \"1138b51b-edf9-42d2-8ea4-4d4b9122eae4\") " pod="openstack/horizon-94dfcc48d-mlrjd" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.643885 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1138b51b-edf9-42d2-8ea4-4d4b9122eae4-horizon-secret-key\") pod \"horizon-94dfcc48d-mlrjd\" (UID: \"1138b51b-edf9-42d2-8ea4-4d4b9122eae4\") " pod="openstack/horizon-94dfcc48d-mlrjd" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.647072 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1138b51b-edf9-42d2-8ea4-4d4b9122eae4-combined-ca-bundle\") pod \"horizon-94dfcc48d-mlrjd\" (UID: \"1138b51b-edf9-42d2-8ea4-4d4b9122eae4\") " pod="openstack/horizon-94dfcc48d-mlrjd" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.658595 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wb2sf\" (UniqueName: \"kubernetes.io/projected/1138b51b-edf9-42d2-8ea4-4d4b9122eae4-kube-api-access-wb2sf\") pod \"horizon-94dfcc48d-mlrjd\" (UID: \"1138b51b-edf9-42d2-8ea4-4d4b9122eae4\") " pod="openstack/horizon-94dfcc48d-mlrjd" Sep 30 14:12:48 crc kubenswrapper[4840]: I0930 14:12:48.730740 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-94dfcc48d-mlrjd" Sep 30 14:12:49 crc kubenswrapper[4840]: I0930 14:12:49.782700 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8b5c85b87-gxf58" Sep 30 14:12:49 crc kubenswrapper[4840]: I0930 14:12:49.839656 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-k8nq4"] Sep 30 14:12:49 crc kubenswrapper[4840]: I0930 14:12:49.839940 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77585f5f8c-k8nq4" podUID="16f65bcb-2335-47af-9705-86e0131a8f7c" containerName="dnsmasq-dns" containerID="cri-o://3e4d847f6209271b5d7feb61da4bc7a0a0745742755f37fd279116bb4f538388" gracePeriod=10 Sep 30 14:12:50 crc kubenswrapper[4840]: I0930 14:12:50.057169 4840 generic.go:334] "Generic (PLEG): container finished" podID="16f65bcb-2335-47af-9705-86e0131a8f7c" containerID="3e4d847f6209271b5d7feb61da4bc7a0a0745742755f37fd279116bb4f538388" exitCode=0 Sep 30 14:12:50 crc kubenswrapper[4840]: I0930 14:12:50.057581 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-k8nq4" event={"ID":"16f65bcb-2335-47af-9705-86e0131a8f7c","Type":"ContainerDied","Data":"3e4d847f6209271b5d7feb61da4bc7a0a0745742755f37fd279116bb4f538388"} Sep 30 14:12:51 crc kubenswrapper[4840]: I0930 14:12:51.871412 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:12:51 crc kubenswrapper[4840]: I0930 14:12:51.871765 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:12:52 crc kubenswrapper[4840]: I0930 14:12:52.331180 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-pxgwt" Sep 30 14:12:52 crc kubenswrapper[4840]: I0930 14:12:52.419516 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1407ce03-21cb-4f1a-9719-6a9aa88a8973-config-data\") pod \"1407ce03-21cb-4f1a-9719-6a9aa88a8973\" (UID: \"1407ce03-21cb-4f1a-9719-6a9aa88a8973\") " Sep 30 14:12:52 crc kubenswrapper[4840]: I0930 14:12:52.419580 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qr7v9\" (UniqueName: \"kubernetes.io/projected/1407ce03-21cb-4f1a-9719-6a9aa88a8973-kube-api-access-qr7v9\") pod \"1407ce03-21cb-4f1a-9719-6a9aa88a8973\" (UID: \"1407ce03-21cb-4f1a-9719-6a9aa88a8973\") " Sep 30 14:12:52 crc kubenswrapper[4840]: I0930 14:12:52.419701 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1407ce03-21cb-4f1a-9719-6a9aa88a8973-combined-ca-bundle\") pod \"1407ce03-21cb-4f1a-9719-6a9aa88a8973\" (UID: \"1407ce03-21cb-4f1a-9719-6a9aa88a8973\") " Sep 30 14:12:52 crc kubenswrapper[4840]: I0930 14:12:52.420039 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1407ce03-21cb-4f1a-9719-6a9aa88a8973-scripts\") pod \"1407ce03-21cb-4f1a-9719-6a9aa88a8973\" (UID: \"1407ce03-21cb-4f1a-9719-6a9aa88a8973\") " Sep 30 14:12:52 crc kubenswrapper[4840]: I0930 14:12:52.420079 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1407ce03-21cb-4f1a-9719-6a9aa88a8973-fernet-keys\") pod \"1407ce03-21cb-4f1a-9719-6a9aa88a8973\" (UID: \"1407ce03-21cb-4f1a-9719-6a9aa88a8973\") " Sep 30 14:12:52 crc kubenswrapper[4840]: I0930 14:12:52.420105 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1407ce03-21cb-4f1a-9719-6a9aa88a8973-credential-keys\") pod \"1407ce03-21cb-4f1a-9719-6a9aa88a8973\" (UID: \"1407ce03-21cb-4f1a-9719-6a9aa88a8973\") " Sep 30 14:12:52 crc kubenswrapper[4840]: I0930 14:12:52.426139 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1407ce03-21cb-4f1a-9719-6a9aa88a8973-kube-api-access-qr7v9" (OuterVolumeSpecName: "kube-api-access-qr7v9") pod "1407ce03-21cb-4f1a-9719-6a9aa88a8973" (UID: "1407ce03-21cb-4f1a-9719-6a9aa88a8973"). InnerVolumeSpecName "kube-api-access-qr7v9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:12:52 crc kubenswrapper[4840]: I0930 14:12:52.427615 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1407ce03-21cb-4f1a-9719-6a9aa88a8973-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "1407ce03-21cb-4f1a-9719-6a9aa88a8973" (UID: "1407ce03-21cb-4f1a-9719-6a9aa88a8973"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:12:52 crc kubenswrapper[4840]: I0930 14:12:52.440632 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1407ce03-21cb-4f1a-9719-6a9aa88a8973-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "1407ce03-21cb-4f1a-9719-6a9aa88a8973" (UID: "1407ce03-21cb-4f1a-9719-6a9aa88a8973"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:12:52 crc kubenswrapper[4840]: I0930 14:12:52.440721 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1407ce03-21cb-4f1a-9719-6a9aa88a8973-scripts" (OuterVolumeSpecName: "scripts") pod "1407ce03-21cb-4f1a-9719-6a9aa88a8973" (UID: "1407ce03-21cb-4f1a-9719-6a9aa88a8973"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:12:52 crc kubenswrapper[4840]: I0930 14:12:52.457866 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1407ce03-21cb-4f1a-9719-6a9aa88a8973-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1407ce03-21cb-4f1a-9719-6a9aa88a8973" (UID: "1407ce03-21cb-4f1a-9719-6a9aa88a8973"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:12:52 crc kubenswrapper[4840]: I0930 14:12:52.462939 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1407ce03-21cb-4f1a-9719-6a9aa88a8973-config-data" (OuterVolumeSpecName: "config-data") pod "1407ce03-21cb-4f1a-9719-6a9aa88a8973" (UID: "1407ce03-21cb-4f1a-9719-6a9aa88a8973"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:12:52 crc kubenswrapper[4840]: I0930 14:12:52.522038 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1407ce03-21cb-4f1a-9719-6a9aa88a8973-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:52 crc kubenswrapper[4840]: I0930 14:12:52.522074 4840 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1407ce03-21cb-4f1a-9719-6a9aa88a8973-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:52 crc kubenswrapper[4840]: I0930 14:12:52.522083 4840 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1407ce03-21cb-4f1a-9719-6a9aa88a8973-fernet-keys\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:52 crc kubenswrapper[4840]: I0930 14:12:52.522092 4840 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1407ce03-21cb-4f1a-9719-6a9aa88a8973-credential-keys\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:52 crc kubenswrapper[4840]: I0930 14:12:52.522100 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1407ce03-21cb-4f1a-9719-6a9aa88a8973-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:52 crc kubenswrapper[4840]: I0930 14:12:52.522108 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qr7v9\" (UniqueName: \"kubernetes.io/projected/1407ce03-21cb-4f1a-9719-6a9aa88a8973-kube-api-access-qr7v9\") on node \"crc\" DevicePath \"\"" Sep 30 14:12:53 crc kubenswrapper[4840]: I0930 14:12:53.082704 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-pxgwt" event={"ID":"1407ce03-21cb-4f1a-9719-6a9aa88a8973","Type":"ContainerDied","Data":"4ddbbee15575c768bf713c18158e98db8ba254d7abc083844f81c3122569692c"} Sep 30 14:12:53 crc kubenswrapper[4840]: I0930 14:12:53.083022 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ddbbee15575c768bf713c18158e98db8ba254d7abc083844f81c3122569692c" Sep 30 14:12:53 crc kubenswrapper[4840]: I0930 14:12:53.082761 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-pxgwt" Sep 30 14:12:53 crc kubenswrapper[4840]: I0930 14:12:53.510504 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-pxgwt"] Sep 30 14:12:53 crc kubenswrapper[4840]: I0930 14:12:53.525345 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-pxgwt"] Sep 30 14:12:53 crc kubenswrapper[4840]: I0930 14:12:53.606183 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-5gzc5"] Sep 30 14:12:53 crc kubenswrapper[4840]: E0930 14:12:53.606907 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1407ce03-21cb-4f1a-9719-6a9aa88a8973" containerName="keystone-bootstrap" Sep 30 14:12:53 crc kubenswrapper[4840]: I0930 14:12:53.606941 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="1407ce03-21cb-4f1a-9719-6a9aa88a8973" containerName="keystone-bootstrap" Sep 30 14:12:53 crc kubenswrapper[4840]: I0930 14:12:53.607398 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="1407ce03-21cb-4f1a-9719-6a9aa88a8973" containerName="keystone-bootstrap" Sep 30 14:12:53 crc kubenswrapper[4840]: I0930 14:12:53.608535 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5gzc5" Sep 30 14:12:53 crc kubenswrapper[4840]: I0930 14:12:53.610778 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Sep 30 14:12:53 crc kubenswrapper[4840]: I0930 14:12:53.610850 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Sep 30 14:12:53 crc kubenswrapper[4840]: I0930 14:12:53.611127 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Sep 30 14:12:53 crc kubenswrapper[4840]: I0930 14:12:53.611639 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gtvd7" Sep 30 14:12:53 crc kubenswrapper[4840]: I0930 14:12:53.619209 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-5gzc5"] Sep 30 14:12:53 crc kubenswrapper[4840]: I0930 14:12:53.743576 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f907c87-9649-4dd2-a419-97055feb4d18-combined-ca-bundle\") pod \"keystone-bootstrap-5gzc5\" (UID: \"0f907c87-9649-4dd2-a419-97055feb4d18\") " pod="openstack/keystone-bootstrap-5gzc5" Sep 30 14:12:53 crc kubenswrapper[4840]: I0930 14:12:53.743664 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f907c87-9649-4dd2-a419-97055feb4d18-scripts\") pod \"keystone-bootstrap-5gzc5\" (UID: \"0f907c87-9649-4dd2-a419-97055feb4d18\") " pod="openstack/keystone-bootstrap-5gzc5" Sep 30 14:12:53 crc kubenswrapper[4840]: I0930 14:12:53.743806 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0f907c87-9649-4dd2-a419-97055feb4d18-credential-keys\") pod \"keystone-bootstrap-5gzc5\" (UID: \"0f907c87-9649-4dd2-a419-97055feb4d18\") " pod="openstack/keystone-bootstrap-5gzc5" Sep 30 14:12:53 crc kubenswrapper[4840]: I0930 14:12:53.743851 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59r2n\" (UniqueName: \"kubernetes.io/projected/0f907c87-9649-4dd2-a419-97055feb4d18-kube-api-access-59r2n\") pod \"keystone-bootstrap-5gzc5\" (UID: \"0f907c87-9649-4dd2-a419-97055feb4d18\") " pod="openstack/keystone-bootstrap-5gzc5" Sep 30 14:12:53 crc kubenswrapper[4840]: I0930 14:12:53.744141 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0f907c87-9649-4dd2-a419-97055feb4d18-fernet-keys\") pod \"keystone-bootstrap-5gzc5\" (UID: \"0f907c87-9649-4dd2-a419-97055feb4d18\") " pod="openstack/keystone-bootstrap-5gzc5" Sep 30 14:12:53 crc kubenswrapper[4840]: I0930 14:12:53.744193 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f907c87-9649-4dd2-a419-97055feb4d18-config-data\") pod \"keystone-bootstrap-5gzc5\" (UID: \"0f907c87-9649-4dd2-a419-97055feb4d18\") " pod="openstack/keystone-bootstrap-5gzc5" Sep 30 14:12:53 crc kubenswrapper[4840]: I0930 14:12:53.846321 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0f907c87-9649-4dd2-a419-97055feb4d18-fernet-keys\") pod \"keystone-bootstrap-5gzc5\" (UID: \"0f907c87-9649-4dd2-a419-97055feb4d18\") " pod="openstack/keystone-bootstrap-5gzc5" Sep 30 14:12:53 crc kubenswrapper[4840]: I0930 14:12:53.846375 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f907c87-9649-4dd2-a419-97055feb4d18-config-data\") pod \"keystone-bootstrap-5gzc5\" (UID: \"0f907c87-9649-4dd2-a419-97055feb4d18\") " pod="openstack/keystone-bootstrap-5gzc5" Sep 30 14:12:53 crc kubenswrapper[4840]: I0930 14:12:53.846457 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f907c87-9649-4dd2-a419-97055feb4d18-combined-ca-bundle\") pod \"keystone-bootstrap-5gzc5\" (UID: \"0f907c87-9649-4dd2-a419-97055feb4d18\") " pod="openstack/keystone-bootstrap-5gzc5" Sep 30 14:12:53 crc kubenswrapper[4840]: I0930 14:12:53.846491 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f907c87-9649-4dd2-a419-97055feb4d18-scripts\") pod \"keystone-bootstrap-5gzc5\" (UID: \"0f907c87-9649-4dd2-a419-97055feb4d18\") " pod="openstack/keystone-bootstrap-5gzc5" Sep 30 14:12:53 crc kubenswrapper[4840]: I0930 14:12:53.846514 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0f907c87-9649-4dd2-a419-97055feb4d18-credential-keys\") pod \"keystone-bootstrap-5gzc5\" (UID: \"0f907c87-9649-4dd2-a419-97055feb4d18\") " pod="openstack/keystone-bootstrap-5gzc5" Sep 30 14:12:53 crc kubenswrapper[4840]: I0930 14:12:53.846533 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59r2n\" (UniqueName: \"kubernetes.io/projected/0f907c87-9649-4dd2-a419-97055feb4d18-kube-api-access-59r2n\") pod \"keystone-bootstrap-5gzc5\" (UID: \"0f907c87-9649-4dd2-a419-97055feb4d18\") " pod="openstack/keystone-bootstrap-5gzc5" Sep 30 14:12:53 crc kubenswrapper[4840]: I0930 14:12:53.851209 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0f907c87-9649-4dd2-a419-97055feb4d18-fernet-keys\") pod \"keystone-bootstrap-5gzc5\" (UID: \"0f907c87-9649-4dd2-a419-97055feb4d18\") " pod="openstack/keystone-bootstrap-5gzc5" Sep 30 14:12:53 crc kubenswrapper[4840]: I0930 14:12:53.851301 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0f907c87-9649-4dd2-a419-97055feb4d18-credential-keys\") pod \"keystone-bootstrap-5gzc5\" (UID: \"0f907c87-9649-4dd2-a419-97055feb4d18\") " pod="openstack/keystone-bootstrap-5gzc5" Sep 30 14:12:53 crc kubenswrapper[4840]: I0930 14:12:53.851767 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f907c87-9649-4dd2-a419-97055feb4d18-scripts\") pod \"keystone-bootstrap-5gzc5\" (UID: \"0f907c87-9649-4dd2-a419-97055feb4d18\") " pod="openstack/keystone-bootstrap-5gzc5" Sep 30 14:12:53 crc kubenswrapper[4840]: I0930 14:12:53.851957 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f907c87-9649-4dd2-a419-97055feb4d18-combined-ca-bundle\") pod \"keystone-bootstrap-5gzc5\" (UID: \"0f907c87-9649-4dd2-a419-97055feb4d18\") " pod="openstack/keystone-bootstrap-5gzc5" Sep 30 14:12:53 crc kubenswrapper[4840]: I0930 14:12:53.853022 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f907c87-9649-4dd2-a419-97055feb4d18-config-data\") pod \"keystone-bootstrap-5gzc5\" (UID: \"0f907c87-9649-4dd2-a419-97055feb4d18\") " pod="openstack/keystone-bootstrap-5gzc5" Sep 30 14:12:53 crc kubenswrapper[4840]: I0930 14:12:53.861891 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59r2n\" (UniqueName: \"kubernetes.io/projected/0f907c87-9649-4dd2-a419-97055feb4d18-kube-api-access-59r2n\") pod \"keystone-bootstrap-5gzc5\" (UID: \"0f907c87-9649-4dd2-a419-97055feb4d18\") " pod="openstack/keystone-bootstrap-5gzc5" Sep 30 14:12:53 crc kubenswrapper[4840]: I0930 14:12:53.934355 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5gzc5" Sep 30 14:12:54 crc kubenswrapper[4840]: I0930 14:12:54.127995 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1407ce03-21cb-4f1a-9719-6a9aa88a8973" path="/var/lib/kubelet/pods/1407ce03-21cb-4f1a-9719-6a9aa88a8973/volumes" Sep 30 14:12:54 crc kubenswrapper[4840]: I0930 14:12:54.392525 4840 scope.go:117] "RemoveContainer" containerID="a2afef429e08db82a843b5ebc0fa79403f00fe910e2d02f574d9c35fdede612f" Sep 30 14:12:54 crc kubenswrapper[4840]: E0930 14:12:54.707430 4840 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Sep 30 14:12:54 crc kubenswrapper[4840]: E0930 14:12:54.707933 4840 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nbfh57ch55chddh564h8h9h67fh649h59h6bhb4h5f6h67h5cdh668h674h597h98h68ch566h6h58dh8h56fh645h5d6h5bh565h64fh677hb7q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hnb5k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(e4a730c4-623d-4ced-a687-bec2d627eae9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 14:12:57 crc kubenswrapper[4840]: I0930 14:12:57.811431 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77585f5f8c-k8nq4" podUID="16f65bcb-2335-47af-9705-86e0131a8f7c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.125:5353: i/o timeout" Sep 30 14:13:02 crc kubenswrapper[4840]: I0930 14:13:02.812243 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77585f5f8c-k8nq4" podUID="16f65bcb-2335-47af-9705-86e0131a8f7c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.125:5353: i/o timeout" Sep 30 14:13:04 crc kubenswrapper[4840]: E0930 14:13:04.137583 4840 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Sep 30 14:13:04 crc kubenswrapper[4840]: E0930 14:13:04.138087 4840 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7g529,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-sp7q5_openstack(ad748b83-4edd-4545-9386-032185c531c5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 14:13:04 crc kubenswrapper[4840]: E0930 14:13:04.139290 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-sp7q5" podUID="ad748b83-4edd-4545-9386-032185c531c5" Sep 30 14:13:04 crc kubenswrapper[4840]: I0930 14:13:04.172533 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-k8nq4" event={"ID":"16f65bcb-2335-47af-9705-86e0131a8f7c","Type":"ContainerDied","Data":"b34ece7c7734d004459e03dac41163ff260ca8475e97f8258b1a1d13671c9f5e"} Sep 30 14:13:04 crc kubenswrapper[4840]: I0930 14:13:04.172602 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b34ece7c7734d004459e03dac41163ff260ca8475e97f8258b1a1d13671c9f5e" Sep 30 14:13:04 crc kubenswrapper[4840]: E0930 14:13:04.175713 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-sp7q5" podUID="ad748b83-4edd-4545-9386-032185c531c5" Sep 30 14:13:04 crc kubenswrapper[4840]: I0930 14:13:04.246996 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-k8nq4" Sep 30 14:13:04 crc kubenswrapper[4840]: I0930 14:13:04.333074 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/16f65bcb-2335-47af-9705-86e0131a8f7c-ovsdbserver-sb\") pod \"16f65bcb-2335-47af-9705-86e0131a8f7c\" (UID: \"16f65bcb-2335-47af-9705-86e0131a8f7c\") " Sep 30 14:13:04 crc kubenswrapper[4840]: I0930 14:13:04.333161 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/16f65bcb-2335-47af-9705-86e0131a8f7c-dns-swift-storage-0\") pod \"16f65bcb-2335-47af-9705-86e0131a8f7c\" (UID: \"16f65bcb-2335-47af-9705-86e0131a8f7c\") " Sep 30 14:13:04 crc kubenswrapper[4840]: I0930 14:13:04.333306 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16f65bcb-2335-47af-9705-86e0131a8f7c-config\") pod \"16f65bcb-2335-47af-9705-86e0131a8f7c\" (UID: \"16f65bcb-2335-47af-9705-86e0131a8f7c\") " Sep 30 14:13:04 crc kubenswrapper[4840]: I0930 14:13:04.333356 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/16f65bcb-2335-47af-9705-86e0131a8f7c-ovsdbserver-nb\") pod \"16f65bcb-2335-47af-9705-86e0131a8f7c\" (UID: \"16f65bcb-2335-47af-9705-86e0131a8f7c\") " Sep 30 14:13:04 crc kubenswrapper[4840]: I0930 14:13:04.333402 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/16f65bcb-2335-47af-9705-86e0131a8f7c-dns-svc\") pod \"16f65bcb-2335-47af-9705-86e0131a8f7c\" (UID: \"16f65bcb-2335-47af-9705-86e0131a8f7c\") " Sep 30 14:13:04 crc kubenswrapper[4840]: I0930 14:13:04.333452 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8hnrx\" (UniqueName: \"kubernetes.io/projected/16f65bcb-2335-47af-9705-86e0131a8f7c-kube-api-access-8hnrx\") pod \"16f65bcb-2335-47af-9705-86e0131a8f7c\" (UID: \"16f65bcb-2335-47af-9705-86e0131a8f7c\") " Sep 30 14:13:04 crc kubenswrapper[4840]: I0930 14:13:04.339854 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16f65bcb-2335-47af-9705-86e0131a8f7c-kube-api-access-8hnrx" (OuterVolumeSpecName: "kube-api-access-8hnrx") pod "16f65bcb-2335-47af-9705-86e0131a8f7c" (UID: "16f65bcb-2335-47af-9705-86e0131a8f7c"). InnerVolumeSpecName "kube-api-access-8hnrx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:13:04 crc kubenswrapper[4840]: I0930 14:13:04.387146 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16f65bcb-2335-47af-9705-86e0131a8f7c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "16f65bcb-2335-47af-9705-86e0131a8f7c" (UID: "16f65bcb-2335-47af-9705-86e0131a8f7c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:13:04 crc kubenswrapper[4840]: I0930 14:13:04.394620 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16f65bcb-2335-47af-9705-86e0131a8f7c-config" (OuterVolumeSpecName: "config") pod "16f65bcb-2335-47af-9705-86e0131a8f7c" (UID: "16f65bcb-2335-47af-9705-86e0131a8f7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:13:04 crc kubenswrapper[4840]: I0930 14:13:04.395422 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16f65bcb-2335-47af-9705-86e0131a8f7c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "16f65bcb-2335-47af-9705-86e0131a8f7c" (UID: "16f65bcb-2335-47af-9705-86e0131a8f7c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:13:04 crc kubenswrapper[4840]: I0930 14:13:04.395614 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16f65bcb-2335-47af-9705-86e0131a8f7c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "16f65bcb-2335-47af-9705-86e0131a8f7c" (UID: "16f65bcb-2335-47af-9705-86e0131a8f7c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:13:04 crc kubenswrapper[4840]: I0930 14:13:04.412496 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16f65bcb-2335-47af-9705-86e0131a8f7c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "16f65bcb-2335-47af-9705-86e0131a8f7c" (UID: "16f65bcb-2335-47af-9705-86e0131a8f7c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:13:04 crc kubenswrapper[4840]: I0930 14:13:04.435384 4840 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16f65bcb-2335-47af-9705-86e0131a8f7c-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:04 crc kubenswrapper[4840]: I0930 14:13:04.435429 4840 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/16f65bcb-2335-47af-9705-86e0131a8f7c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:04 crc kubenswrapper[4840]: I0930 14:13:04.435442 4840 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/16f65bcb-2335-47af-9705-86e0131a8f7c-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:04 crc kubenswrapper[4840]: I0930 14:13:04.435453 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8hnrx\" (UniqueName: \"kubernetes.io/projected/16f65bcb-2335-47af-9705-86e0131a8f7c-kube-api-access-8hnrx\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:04 crc kubenswrapper[4840]: I0930 14:13:04.435465 4840 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/16f65bcb-2335-47af-9705-86e0131a8f7c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:04 crc kubenswrapper[4840]: I0930 14:13:04.435477 4840 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/16f65bcb-2335-47af-9705-86e0131a8f7c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:04 crc kubenswrapper[4840]: E0930 14:13:04.658347 4840 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Sep 30 14:13:04 crc kubenswrapper[4840]: E0930 14:13:04.658770 4840 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4l6jg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-9hcwn_openstack(2c465939-fcbe-44b2-b7ae-a24b5a581dbc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 14:13:04 crc kubenswrapper[4840]: E0930 14:13:04.659945 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-9hcwn" podUID="2c465939-fcbe-44b2-b7ae-a24b5a581dbc" Sep 30 14:13:05 crc kubenswrapper[4840]: I0930 14:13:05.186264 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-k8nq4" Sep 30 14:13:05 crc kubenswrapper[4840]: E0930 14:13:05.188525 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-9hcwn" podUID="2c465939-fcbe-44b2-b7ae-a24b5a581dbc" Sep 30 14:13:05 crc kubenswrapper[4840]: I0930 14:13:05.229853 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-k8nq4"] Sep 30 14:13:05 crc kubenswrapper[4840]: I0930 14:13:05.236019 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-k8nq4"] Sep 30 14:13:05 crc kubenswrapper[4840]: E0930 14:13:05.377946 4840 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod16f65bcb_2335_47af_9705_86e0131a8f7c.slice/crio-b34ece7c7734d004459e03dac41163ff260ca8475e97f8258b1a1d13671c9f5e\": RecentStats: unable to find data in memory cache]" Sep 30 14:13:05 crc kubenswrapper[4840]: I0930 14:13:05.650485 4840 scope.go:117] "RemoveContainer" containerID="6898fce6551a8735dee36a89b338612548caef2e3441fd6383b3783ca715ddfc" Sep 30 14:13:05 crc kubenswrapper[4840]: E0930 14:13:05.651430 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6898fce6551a8735dee36a89b338612548caef2e3441fd6383b3783ca715ddfc\": container with ID starting with 6898fce6551a8735dee36a89b338612548caef2e3441fd6383b3783ca715ddfc not found: ID does not exist" containerID="6898fce6551a8735dee36a89b338612548caef2e3441fd6383b3783ca715ddfc" Sep 30 14:13:05 crc kubenswrapper[4840]: I0930 14:13:05.651482 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6898fce6551a8735dee36a89b338612548caef2e3441fd6383b3783ca715ddfc"} err="failed to get container status \"6898fce6551a8735dee36a89b338612548caef2e3441fd6383b3783ca715ddfc\": rpc error: code = NotFound desc = could not find container \"6898fce6551a8735dee36a89b338612548caef2e3441fd6383b3783ca715ddfc\": container with ID starting with 6898fce6551a8735dee36a89b338612548caef2e3441fd6383b3783ca715ddfc not found: ID does not exist" Sep 30 14:13:05 crc kubenswrapper[4840]: I0930 14:13:05.651520 4840 scope.go:117] "RemoveContainer" containerID="a2afef429e08db82a843b5ebc0fa79403f00fe910e2d02f574d9c35fdede612f" Sep 30 14:13:05 crc kubenswrapper[4840]: E0930 14:13:05.651896 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2afef429e08db82a843b5ebc0fa79403f00fe910e2d02f574d9c35fdede612f\": container with ID starting with a2afef429e08db82a843b5ebc0fa79403f00fe910e2d02f574d9c35fdede612f not found: ID does not exist" containerID="a2afef429e08db82a843b5ebc0fa79403f00fe910e2d02f574d9c35fdede612f" Sep 30 14:13:05 crc kubenswrapper[4840]: I0930 14:13:05.651923 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2afef429e08db82a843b5ebc0fa79403f00fe910e2d02f574d9c35fdede612f"} err="failed to get container status \"a2afef429e08db82a843b5ebc0fa79403f00fe910e2d02f574d9c35fdede612f\": rpc error: code = NotFound desc = could not find container \"a2afef429e08db82a843b5ebc0fa79403f00fe910e2d02f574d9c35fdede612f\": container with ID starting with a2afef429e08db82a843b5ebc0fa79403f00fe910e2d02f574d9c35fdede612f not found: ID does not exist" Sep 30 14:13:05 crc kubenswrapper[4840]: I0930 14:13:05.651941 4840 scope.go:117] "RemoveContainer" containerID="6898fce6551a8735dee36a89b338612548caef2e3441fd6383b3783ca715ddfc" Sep 30 14:13:05 crc kubenswrapper[4840]: I0930 14:13:05.653033 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6898fce6551a8735dee36a89b338612548caef2e3441fd6383b3783ca715ddfc"} err="failed to get container status \"6898fce6551a8735dee36a89b338612548caef2e3441fd6383b3783ca715ddfc\": rpc error: code = NotFound desc = could not find container \"6898fce6551a8735dee36a89b338612548caef2e3441fd6383b3783ca715ddfc\": container with ID starting with 6898fce6551a8735dee36a89b338612548caef2e3441fd6383b3783ca715ddfc not found: ID does not exist" Sep 30 14:13:05 crc kubenswrapper[4840]: I0930 14:13:05.653090 4840 scope.go:117] "RemoveContainer" containerID="a2afef429e08db82a843b5ebc0fa79403f00fe910e2d02f574d9c35fdede612f" Sep 30 14:13:05 crc kubenswrapper[4840]: I0930 14:13:05.653529 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2afef429e08db82a843b5ebc0fa79403f00fe910e2d02f574d9c35fdede612f"} err="failed to get container status \"a2afef429e08db82a843b5ebc0fa79403f00fe910e2d02f574d9c35fdede612f\": rpc error: code = NotFound desc = could not find container \"a2afef429e08db82a843b5ebc0fa79403f00fe910e2d02f574d9c35fdede612f\": container with ID starting with a2afef429e08db82a843b5ebc0fa79403f00fe910e2d02f574d9c35fdede612f not found: ID does not exist" Sep 30 14:13:05 crc kubenswrapper[4840]: I0930 14:13:05.653576 4840 scope.go:117] "RemoveContainer" containerID="70029353c5195b1b6f43700b02ee9bc10d4d6aaf1ea4ac7ca79015b811c337eb" Sep 30 14:13:05 crc kubenswrapper[4840]: E0930 14:13:05.681372 4840 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Sep 30 14:13:05 crc kubenswrapper[4840]: E0930 14:13:05.681509 4840 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pjbfr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-9vd76_openstack(242d13bb-6113-421b-83d1-fcccf91a9c68): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 14:13:05 crc kubenswrapper[4840]: E0930 14:13:05.682778 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-9vd76" podUID="242d13bb-6113-421b-83d1-fcccf91a9c68" Sep 30 14:13:05 crc kubenswrapper[4840]: I0930 14:13:05.944175 4840 scope.go:117] "RemoveContainer" containerID="7aa98fc6015e602ca6609d69a712c47542cdb3a0afd32ad496cfb5db2098b7a7" Sep 30 14:13:06 crc kubenswrapper[4840]: I0930 14:13:06.067807 4840 scope.go:117] "RemoveContainer" containerID="70029353c5195b1b6f43700b02ee9bc10d4d6aaf1ea4ac7ca79015b811c337eb" Sep 30 14:13:06 crc kubenswrapper[4840]: E0930 14:13:06.068581 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70029353c5195b1b6f43700b02ee9bc10d4d6aaf1ea4ac7ca79015b811c337eb\": container with ID starting with 70029353c5195b1b6f43700b02ee9bc10d4d6aaf1ea4ac7ca79015b811c337eb not found: ID does not exist" containerID="70029353c5195b1b6f43700b02ee9bc10d4d6aaf1ea4ac7ca79015b811c337eb" Sep 30 14:13:06 crc kubenswrapper[4840]: I0930 14:13:06.068624 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70029353c5195b1b6f43700b02ee9bc10d4d6aaf1ea4ac7ca79015b811c337eb"} err="failed to get container status \"70029353c5195b1b6f43700b02ee9bc10d4d6aaf1ea4ac7ca79015b811c337eb\": rpc error: code = NotFound desc = could not find container \"70029353c5195b1b6f43700b02ee9bc10d4d6aaf1ea4ac7ca79015b811c337eb\": container with ID starting with 70029353c5195b1b6f43700b02ee9bc10d4d6aaf1ea4ac7ca79015b811c337eb not found: ID does not exist" Sep 30 14:13:06 crc kubenswrapper[4840]: I0930 14:13:06.068651 4840 scope.go:117] "RemoveContainer" containerID="7aa98fc6015e602ca6609d69a712c47542cdb3a0afd32ad496cfb5db2098b7a7" Sep 30 14:13:06 crc kubenswrapper[4840]: E0930 14:13:06.070252 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7aa98fc6015e602ca6609d69a712c47542cdb3a0afd32ad496cfb5db2098b7a7\": container with ID starting with 7aa98fc6015e602ca6609d69a712c47542cdb3a0afd32ad496cfb5db2098b7a7 not found: ID does not exist" containerID="7aa98fc6015e602ca6609d69a712c47542cdb3a0afd32ad496cfb5db2098b7a7" Sep 30 14:13:06 crc kubenswrapper[4840]: I0930 14:13:06.070277 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7aa98fc6015e602ca6609d69a712c47542cdb3a0afd32ad496cfb5db2098b7a7"} err="failed to get container status \"7aa98fc6015e602ca6609d69a712c47542cdb3a0afd32ad496cfb5db2098b7a7\": rpc error: code = NotFound desc = could not find container \"7aa98fc6015e602ca6609d69a712c47542cdb3a0afd32ad496cfb5db2098b7a7\": container with ID starting with 7aa98fc6015e602ca6609d69a712c47542cdb3a0afd32ad496cfb5db2098b7a7 not found: ID does not exist" Sep 30 14:13:06 crc kubenswrapper[4840]: I0930 14:13:06.070294 4840 scope.go:117] "RemoveContainer" containerID="70029353c5195b1b6f43700b02ee9bc10d4d6aaf1ea4ac7ca79015b811c337eb" Sep 30 14:13:06 crc kubenswrapper[4840]: I0930 14:13:06.070667 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70029353c5195b1b6f43700b02ee9bc10d4d6aaf1ea4ac7ca79015b811c337eb"} err="failed to get container status \"70029353c5195b1b6f43700b02ee9bc10d4d6aaf1ea4ac7ca79015b811c337eb\": rpc error: code = NotFound desc = could not find container \"70029353c5195b1b6f43700b02ee9bc10d4d6aaf1ea4ac7ca79015b811c337eb\": container with ID starting with 70029353c5195b1b6f43700b02ee9bc10d4d6aaf1ea4ac7ca79015b811c337eb not found: ID does not exist" Sep 30 14:13:06 crc kubenswrapper[4840]: I0930 14:13:06.070713 4840 scope.go:117] "RemoveContainer" containerID="7aa98fc6015e602ca6609d69a712c47542cdb3a0afd32ad496cfb5db2098b7a7" Sep 30 14:13:06 crc kubenswrapper[4840]: I0930 14:13:06.071406 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7aa98fc6015e602ca6609d69a712c47542cdb3a0afd32ad496cfb5db2098b7a7"} err="failed to get container status \"7aa98fc6015e602ca6609d69a712c47542cdb3a0afd32ad496cfb5db2098b7a7\": rpc error: code = NotFound desc = could not find container \"7aa98fc6015e602ca6609d69a712c47542cdb3a0afd32ad496cfb5db2098b7a7\": container with ID starting with 7aa98fc6015e602ca6609d69a712c47542cdb3a0afd32ad496cfb5db2098b7a7 not found: ID does not exist" Sep 30 14:13:06 crc kubenswrapper[4840]: I0930 14:13:06.133479 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16f65bcb-2335-47af-9705-86e0131a8f7c" path="/var/lib/kubelet/pods/16f65bcb-2335-47af-9705-86e0131a8f7c/volumes" Sep 30 14:13:06 crc kubenswrapper[4840]: I0930 14:13:06.201316 4840 generic.go:334] "Generic (PLEG): container finished" podID="85509efb-abc4-4649-a320-7b9de5487180" containerID="f4c754f0714009dfde0d74882248a19e96c17147f4c8a912359797cef63027ca" exitCode=0 Sep 30 14:13:06 crc kubenswrapper[4840]: I0930 14:13:06.201382 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-nplct" event={"ID":"85509efb-abc4-4649-a320-7b9de5487180","Type":"ContainerDied","Data":"f4c754f0714009dfde0d74882248a19e96c17147f4c8a912359797cef63027ca"} Sep 30 14:13:06 crc kubenswrapper[4840]: E0930 14:13:06.208697 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-9vd76" podUID="242d13bb-6113-421b-83d1-fcccf91a9c68" Sep 30 14:13:06 crc kubenswrapper[4840]: I0930 14:13:06.299855 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5595d79c68-dgsth"] Sep 30 14:13:06 crc kubenswrapper[4840]: W0930 14:13:06.313955 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e9b87b2_4836_4857_8514_b2dcfe4de1f5.slice/crio-9b0c0ec11411cdf62ef15443f9485d30baa67c39b99c6937c92ff75d56b07dd0 WatchSource:0}: Error finding container 9b0c0ec11411cdf62ef15443f9485d30baa67c39b99c6937c92ff75d56b07dd0: Status 404 returned error can't find the container with id 9b0c0ec11411cdf62ef15443f9485d30baa67c39b99c6937c92ff75d56b07dd0 Sep 30 14:13:06 crc kubenswrapper[4840]: I0930 14:13:06.415243 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:13:06 crc kubenswrapper[4840]: W0930 14:13:06.422787 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda7b2bbcc_8c5b_4eaa_ab08_87a738b4adaa.slice/crio-ed8ad838a3d1aa034d4a67b2b948166a2bb568d0ef615fbdd6094de64d36d483 WatchSource:0}: Error finding container ed8ad838a3d1aa034d4a67b2b948166a2bb568d0ef615fbdd6094de64d36d483: Status 404 returned error can't find the container with id ed8ad838a3d1aa034d4a67b2b948166a2bb568d0ef615fbdd6094de64d36d483 Sep 30 14:13:06 crc kubenswrapper[4840]: I0930 14:13:06.451036 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-94dfcc48d-mlrjd"] Sep 30 14:13:06 crc kubenswrapper[4840]: W0930 14:13:06.451126 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1138b51b_edf9_42d2_8ea4_4d4b9122eae4.slice/crio-ed852f7e6bdd8be52b0d9b7b1ea06fcfcc7ff89c016c9f824f4c4ce9919eec5a WatchSource:0}: Error finding container ed852f7e6bdd8be52b0d9b7b1ea06fcfcc7ff89c016c9f824f4c4ce9919eec5a: Status 404 returned error can't find the container with id ed852f7e6bdd8be52b0d9b7b1ea06fcfcc7ff89c016c9f824f4c4ce9919eec5a Sep 30 14:13:06 crc kubenswrapper[4840]: I0930 14:13:06.524534 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-5gzc5"] Sep 30 14:13:06 crc kubenswrapper[4840]: W0930 14:13:06.528400 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f907c87_9649_4dd2_a419_97055feb4d18.slice/crio-f3c9b387244c00a461e6391af9513520d1ff8adb2a889327390ea106c5a648b0 WatchSource:0}: Error finding container f3c9b387244c00a461e6391af9513520d1ff8adb2a889327390ea106c5a648b0: Status 404 returned error can't find the container with id f3c9b387244c00a461e6391af9513520d1ff8adb2a889327390ea106c5a648b0 Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.223111 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e4a730c4-623d-4ced-a687-bec2d627eae9","Type":"ContainerStarted","Data":"19c1bb2cbbc34614446d73b508c35aafb815b7a569fbfdb49916e9cdaadb3519"} Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.229613 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5gzc5" event={"ID":"0f907c87-9649-4dd2-a419-97055feb4d18","Type":"ContainerStarted","Data":"25f1bac61cdc7bb81eac30ec24cc17df033b36b1740140f97751b58d10a731b7"} Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.229662 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5gzc5" event={"ID":"0f907c87-9649-4dd2-a419-97055feb4d18","Type":"ContainerStarted","Data":"f3c9b387244c00a461e6391af9513520d1ff8adb2a889327390ea106c5a648b0"} Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.237469 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-94dfcc48d-mlrjd" event={"ID":"1138b51b-edf9-42d2-8ea4-4d4b9122eae4","Type":"ContainerStarted","Data":"6c22a094756cb0f091e9471f18bd5f5849c9d70d23c79702dd2caee923ea44aa"} Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.237723 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-94dfcc48d-mlrjd" event={"ID":"1138b51b-edf9-42d2-8ea4-4d4b9122eae4","Type":"ContainerStarted","Data":"f692a86c1010f5728fafa6369b5832808ca72898946cdaa5774602c0f1611813"} Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.237740 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-94dfcc48d-mlrjd" event={"ID":"1138b51b-edf9-42d2-8ea4-4d4b9122eae4","Type":"ContainerStarted","Data":"ed852f7e6bdd8be52b0d9b7b1ea06fcfcc7ff89c016c9f824f4c4ce9919eec5a"} Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.247567 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-788d88bd6f-l44pw" event={"ID":"e141523f-ce9b-4626-b155-b0a82a3f6dc6","Type":"ContainerStarted","Data":"71b465222c8ca2c7063b1632b64ef4d2b3dbe73248c1b00977324af0871011fc"} Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.247625 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-788d88bd6f-l44pw" event={"ID":"e141523f-ce9b-4626-b155-b0a82a3f6dc6","Type":"ContainerStarted","Data":"91f9f1d766fa7013e2cf435e7a46c86870d1a985c1f7db12ac2763f63a8b3730"} Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.247790 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-788d88bd6f-l44pw" podUID="e141523f-ce9b-4626-b155-b0a82a3f6dc6" containerName="horizon-log" containerID="cri-o://91f9f1d766fa7013e2cf435e7a46c86870d1a985c1f7db12ac2763f63a8b3730" gracePeriod=30 Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.248122 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-788d88bd6f-l44pw" podUID="e141523f-ce9b-4626-b155-b0a82a3f6dc6" containerName="horizon" containerID="cri-o://71b465222c8ca2c7063b1632b64ef4d2b3dbe73248c1b00977324af0871011fc" gracePeriod=30 Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.258047 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-5gzc5" podStartSLOduration=14.258027315 podStartE2EDuration="14.258027315s" podCreationTimestamp="2025-09-30 14:12:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:13:07.252072 +0000 UTC m=+1015.881158443" watchObservedRunningTime="2025-09-30 14:13:07.258027315 +0000 UTC m=+1015.887113728" Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.263340 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f785c99fc-fpf4g" event={"ID":"47067ffa-fc01-48b4-abcd-4405053eadca","Type":"ContainerStarted","Data":"6d38e955431f3f93cdb79f42366f8efd916449357858ee1e88db0ae9f7808b40"} Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.263402 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f785c99fc-fpf4g" event={"ID":"47067ffa-fc01-48b4-abcd-4405053eadca","Type":"ContainerStarted","Data":"1e4111d6b9033739a506b5a02fccfc6924f5a5fd075ba04c56d92cc8bc086c9e"} Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.263684 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7f785c99fc-fpf4g" podUID="47067ffa-fc01-48b4-abcd-4405053eadca" containerName="horizon-log" containerID="cri-o://1e4111d6b9033739a506b5a02fccfc6924f5a5fd075ba04c56d92cc8bc086c9e" gracePeriod=30 Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.264050 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7f785c99fc-fpf4g" podUID="47067ffa-fc01-48b4-abcd-4405053eadca" containerName="horizon" containerID="cri-o://6d38e955431f3f93cdb79f42366f8efd916449357858ee1e88db0ae9f7808b40" gracePeriod=30 Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.272133 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa","Type":"ContainerStarted","Data":"8150ffc2faf4a5d465566a1a3e0e9f2d18a03857ffcd0eba0fdac856a4125a9b"} Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.272330 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa","Type":"ContainerStarted","Data":"ed8ad838a3d1aa034d4a67b2b948166a2bb568d0ef615fbdd6094de64d36d483"} Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.274314 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-94dfcc48d-mlrjd" podStartSLOduration=19.27430087 podStartE2EDuration="19.27430087s" podCreationTimestamp="2025-09-30 14:12:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:13:07.273432727 +0000 UTC m=+1015.902519170" watchObservedRunningTime="2025-09-30 14:13:07.27430087 +0000 UTC m=+1015.903387293" Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.296306 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5cc547b489-2p58b" podUID="a7f0b012-f515-434f-919e-5f0409c8e554" containerName="horizon-log" containerID="cri-o://807dbc9df37357f57e1252dc8f83d1de62c3336b8cb4150169bcddd5321ae0e3" gracePeriod=30 Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.296424 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5cc547b489-2p58b" podUID="a7f0b012-f515-434f-919e-5f0409c8e554" containerName="horizon" containerID="cri-o://2edee90f84eb32c953eeb5dfa1fc2729836da354666c490c7c4bac46e34c4409" gracePeriod=30 Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.296256 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5cc547b489-2p58b" event={"ID":"a7f0b012-f515-434f-919e-5f0409c8e554","Type":"ContainerStarted","Data":"2edee90f84eb32c953eeb5dfa1fc2729836da354666c490c7c4bac46e34c4409"} Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.296531 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5cc547b489-2p58b" event={"ID":"a7f0b012-f515-434f-919e-5f0409c8e554","Type":"ContainerStarted","Data":"807dbc9df37357f57e1252dc8f83d1de62c3336b8cb4150169bcddd5321ae0e3"} Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.310890 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5595d79c68-dgsth" event={"ID":"5e9b87b2-4836-4857-8514-b2dcfe4de1f5","Type":"ContainerStarted","Data":"728e90849e942f659a1349a083627796a9122008fae88346ec1107d99eefbab1"} Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.310929 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5595d79c68-dgsth" event={"ID":"5e9b87b2-4836-4857-8514-b2dcfe4de1f5","Type":"ContainerStarted","Data":"019b8bf8dc37878bad1354f6206205178cb6737805a6453683b754b0a592f5d1"} Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.310939 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5595d79c68-dgsth" event={"ID":"5e9b87b2-4836-4857-8514-b2dcfe4de1f5","Type":"ContainerStarted","Data":"9b0c0ec11411cdf62ef15443f9485d30baa67c39b99c6937c92ff75d56b07dd0"} Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.310947 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-788d88bd6f-l44pw" podStartSLOduration=3.561628179 podStartE2EDuration="26.310924905s" podCreationTimestamp="2025-09-30 14:12:41 +0000 UTC" firstStartedPulling="2025-09-30 14:12:43.152282295 +0000 UTC m=+991.781368728" lastFinishedPulling="2025-09-30 14:13:05.901579031 +0000 UTC m=+1014.530665454" observedRunningTime="2025-09-30 14:13:07.305189245 +0000 UTC m=+1015.934275668" watchObservedRunningTime="2025-09-30 14:13:07.310924905 +0000 UTC m=+1015.940011328" Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.334905 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5cc547b489-2p58b" podStartSLOduration=3.119933463 podStartE2EDuration="28.33488629s" podCreationTimestamp="2025-09-30 14:12:39 +0000 UTC" firstStartedPulling="2025-09-30 14:12:40.410364361 +0000 UTC m=+989.039450784" lastFinishedPulling="2025-09-30 14:13:05.625317188 +0000 UTC m=+1014.254403611" observedRunningTime="2025-09-30 14:13:07.327778764 +0000 UTC m=+1015.956865187" watchObservedRunningTime="2025-09-30 14:13:07.33488629 +0000 UTC m=+1015.963972713" Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.362512 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7f785c99fc-fpf4g" podStartSLOduration=3.148518348 podStartE2EDuration="28.362485619s" podCreationTimestamp="2025-09-30 14:12:39 +0000 UTC" firstStartedPulling="2025-09-30 14:12:40.68722338 +0000 UTC m=+989.316309803" lastFinishedPulling="2025-09-30 14:13:05.901190651 +0000 UTC m=+1014.530277074" observedRunningTime="2025-09-30 14:13:07.343800612 +0000 UTC m=+1015.972887065" watchObservedRunningTime="2025-09-30 14:13:07.362485619 +0000 UTC m=+1015.991572032" Sep 30 14:13:07 crc kubenswrapper[4840]: W0930 14:13:07.386770 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod91a0f20e_5a31_42fc_ba4e_c5abd44d497c.slice/crio-669f521351fd5967336af154cee4ccd531f6c3667cecc3c32708eb943d74318f WatchSource:0}: Error finding container 669f521351fd5967336af154cee4ccd531f6c3667cecc3c32708eb943d74318f: Status 404 returned error can't find the container with id 669f521351fd5967336af154cee4ccd531f6c3667cecc3c32708eb943d74318f Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.396483 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.401517 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5595d79c68-dgsth" podStartSLOduration=19.401493197 podStartE2EDuration="19.401493197s" podCreationTimestamp="2025-09-30 14:12:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:13:07.369454621 +0000 UTC m=+1015.998541064" watchObservedRunningTime="2025-09-30 14:13:07.401493197 +0000 UTC m=+1016.030579620" Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.771530 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-nplct" Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.803134 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/85509efb-abc4-4649-a320-7b9de5487180-config\") pod \"85509efb-abc4-4649-a320-7b9de5487180\" (UID: \"85509efb-abc4-4649-a320-7b9de5487180\") " Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.803222 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85509efb-abc4-4649-a320-7b9de5487180-combined-ca-bundle\") pod \"85509efb-abc4-4649-a320-7b9de5487180\" (UID: \"85509efb-abc4-4649-a320-7b9de5487180\") " Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.803242 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wc5q2\" (UniqueName: \"kubernetes.io/projected/85509efb-abc4-4649-a320-7b9de5487180-kube-api-access-wc5q2\") pod \"85509efb-abc4-4649-a320-7b9de5487180\" (UID: \"85509efb-abc4-4649-a320-7b9de5487180\") " Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.810247 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85509efb-abc4-4649-a320-7b9de5487180-kube-api-access-wc5q2" (OuterVolumeSpecName: "kube-api-access-wc5q2") pod "85509efb-abc4-4649-a320-7b9de5487180" (UID: "85509efb-abc4-4649-a320-7b9de5487180"). InnerVolumeSpecName "kube-api-access-wc5q2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.813659 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77585f5f8c-k8nq4" podUID="16f65bcb-2335-47af-9705-86e0131a8f7c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.125:5353: i/o timeout" Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.888226 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85509efb-abc4-4649-a320-7b9de5487180-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "85509efb-abc4-4649-a320-7b9de5487180" (UID: "85509efb-abc4-4649-a320-7b9de5487180"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.905242 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85509efb-abc4-4649-a320-7b9de5487180-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.905434 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wc5q2\" (UniqueName: \"kubernetes.io/projected/85509efb-abc4-4649-a320-7b9de5487180-kube-api-access-wc5q2\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:07 crc kubenswrapper[4840]: I0930 14:13:07.907688 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85509efb-abc4-4649-a320-7b9de5487180-config" (OuterVolumeSpecName: "config") pod "85509efb-abc4-4649-a320-7b9de5487180" (UID: "85509efb-abc4-4649-a320-7b9de5487180"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.007175 4840 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/85509efb-abc4-4649-a320-7b9de5487180-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.329041 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa","Type":"ContainerStarted","Data":"801550a13d2831db01a9f0874cc571e7749f7e3f4095df0690b2f7861f87a377"} Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.329185 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa" containerName="glance-log" containerID="cri-o://8150ffc2faf4a5d465566a1a3e0e9f2d18a03857ffcd0eba0fdac856a4125a9b" gracePeriod=30 Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.329624 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa" containerName="glance-httpd" containerID="cri-o://801550a13d2831db01a9f0874cc571e7749f7e3f4095df0690b2f7861f87a377" gracePeriod=30 Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.340748 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-nplct" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.341054 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-nplct" event={"ID":"85509efb-abc4-4649-a320-7b9de5487180","Type":"ContainerDied","Data":"23dcba2ba3c0ac8d8de12a191bf04af09ebebac19e5a6356c0486cd7fc9d4101"} Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.341094 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23dcba2ba3c0ac8d8de12a191bf04af09ebebac19e5a6356c0486cd7fc9d4101" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.346202 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"91a0f20e-5a31-42fc-ba4e-c5abd44d497c","Type":"ContainerStarted","Data":"cdb11879b0067fe94fa10997b5f54c42a9263b63162c6e54eb8ee403ec65cd57"} Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.346241 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"91a0f20e-5a31-42fc-ba4e-c5abd44d497c","Type":"ContainerStarted","Data":"669f521351fd5967336af154cee4ccd531f6c3667cecc3c32708eb943d74318f"} Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.359875 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=23.359857239 podStartE2EDuration="23.359857239s" podCreationTimestamp="2025-09-30 14:12:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:13:08.356264085 +0000 UTC m=+1016.985350508" watchObservedRunningTime="2025-09-30 14:13:08.359857239 +0000 UTC m=+1016.988943652" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.469017 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-zvfqz"] Sep 30 14:13:08 crc kubenswrapper[4840]: E0930 14:13:08.469774 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16f65bcb-2335-47af-9705-86e0131a8f7c" containerName="dnsmasq-dns" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.469797 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="16f65bcb-2335-47af-9705-86e0131a8f7c" containerName="dnsmasq-dns" Sep 30 14:13:08 crc kubenswrapper[4840]: E0930 14:13:08.469809 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85509efb-abc4-4649-a320-7b9de5487180" containerName="neutron-db-sync" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.469816 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="85509efb-abc4-4649-a320-7b9de5487180" containerName="neutron-db-sync" Sep 30 14:13:08 crc kubenswrapper[4840]: E0930 14:13:08.469844 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16f65bcb-2335-47af-9705-86e0131a8f7c" containerName="init" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.469851 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="16f65bcb-2335-47af-9705-86e0131a8f7c" containerName="init" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.470018 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="16f65bcb-2335-47af-9705-86e0131a8f7c" containerName="dnsmasq-dns" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.470042 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="85509efb-abc4-4649-a320-7b9de5487180" containerName="neutron-db-sync" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.475938 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-zvfqz" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.487257 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-zvfqz"] Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.518658 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5c9be753-da64-4599-b5d1-baf7f9be58d4-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-zvfqz\" (UID: \"5c9be753-da64-4599-b5d1-baf7f9be58d4\") " pod="openstack/dnsmasq-dns-84b966f6c9-zvfqz" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.518767 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c9be753-da64-4599-b5d1-baf7f9be58d4-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-zvfqz\" (UID: \"5c9be753-da64-4599-b5d1-baf7f9be58d4\") " pod="openstack/dnsmasq-dns-84b966f6c9-zvfqz" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.518793 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5c9be753-da64-4599-b5d1-baf7f9be58d4-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-zvfqz\" (UID: \"5c9be753-da64-4599-b5d1-baf7f9be58d4\") " pod="openstack/dnsmasq-dns-84b966f6c9-zvfqz" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.518888 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5c9be753-da64-4599-b5d1-baf7f9be58d4-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-zvfqz\" (UID: \"5c9be753-da64-4599-b5d1-baf7f9be58d4\") " pod="openstack/dnsmasq-dns-84b966f6c9-zvfqz" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.518937 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c9be753-da64-4599-b5d1-baf7f9be58d4-config\") pod \"dnsmasq-dns-84b966f6c9-zvfqz\" (UID: \"5c9be753-da64-4599-b5d1-baf7f9be58d4\") " pod="openstack/dnsmasq-dns-84b966f6c9-zvfqz" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.518967 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vt6lb\" (UniqueName: \"kubernetes.io/projected/5c9be753-da64-4599-b5d1-baf7f9be58d4-kube-api-access-vt6lb\") pod \"dnsmasq-dns-84b966f6c9-zvfqz\" (UID: \"5c9be753-da64-4599-b5d1-baf7f9be58d4\") " pod="openstack/dnsmasq-dns-84b966f6c9-zvfqz" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.593792 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5fbb8d5d4d-g2v89"] Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.595691 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5595d79c68-dgsth" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.595746 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5595d79c68-dgsth" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.595895 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5fbb8d5d4d-g2v89" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.599850 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.600464 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.600692 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-7h4wd" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.600873 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.607004 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5fbb8d5d4d-g2v89"] Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.621014 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c9be753-da64-4599-b5d1-baf7f9be58d4-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-zvfqz\" (UID: \"5c9be753-da64-4599-b5d1-baf7f9be58d4\") " pod="openstack/dnsmasq-dns-84b966f6c9-zvfqz" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.621252 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5c9be753-da64-4599-b5d1-baf7f9be58d4-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-zvfqz\" (UID: \"5c9be753-da64-4599-b5d1-baf7f9be58d4\") " pod="openstack/dnsmasq-dns-84b966f6c9-zvfqz" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.621416 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a51d9a5-67ae-488d-a1e7-003d7db4267f-ovndb-tls-certs\") pod \"neutron-5fbb8d5d4d-g2v89\" (UID: \"9a51d9a5-67ae-488d-a1e7-003d7db4267f\") " pod="openstack/neutron-5fbb8d5d4d-g2v89" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.621677 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zclvs\" (UniqueName: \"kubernetes.io/projected/9a51d9a5-67ae-488d-a1e7-003d7db4267f-kube-api-access-zclvs\") pod \"neutron-5fbb8d5d4d-g2v89\" (UID: \"9a51d9a5-67ae-488d-a1e7-003d7db4267f\") " pod="openstack/neutron-5fbb8d5d4d-g2v89" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.621846 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5c9be753-da64-4599-b5d1-baf7f9be58d4-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-zvfqz\" (UID: \"5c9be753-da64-4599-b5d1-baf7f9be58d4\") " pod="openstack/dnsmasq-dns-84b966f6c9-zvfqz" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.621982 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c9be753-da64-4599-b5d1-baf7f9be58d4-config\") pod \"dnsmasq-dns-84b966f6c9-zvfqz\" (UID: \"5c9be753-da64-4599-b5d1-baf7f9be58d4\") " pod="openstack/dnsmasq-dns-84b966f6c9-zvfqz" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.622103 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a51d9a5-67ae-488d-a1e7-003d7db4267f-combined-ca-bundle\") pod \"neutron-5fbb8d5d4d-g2v89\" (UID: \"9a51d9a5-67ae-488d-a1e7-003d7db4267f\") " pod="openstack/neutron-5fbb8d5d4d-g2v89" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.622207 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vt6lb\" (UniqueName: \"kubernetes.io/projected/5c9be753-da64-4599-b5d1-baf7f9be58d4-kube-api-access-vt6lb\") pod \"dnsmasq-dns-84b966f6c9-zvfqz\" (UID: \"5c9be753-da64-4599-b5d1-baf7f9be58d4\") " pod="openstack/dnsmasq-dns-84b966f6c9-zvfqz" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.622354 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5c9be753-da64-4599-b5d1-baf7f9be58d4-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-zvfqz\" (UID: \"5c9be753-da64-4599-b5d1-baf7f9be58d4\") " pod="openstack/dnsmasq-dns-84b966f6c9-zvfqz" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.622471 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9a51d9a5-67ae-488d-a1e7-003d7db4267f-httpd-config\") pod \"neutron-5fbb8d5d4d-g2v89\" (UID: \"9a51d9a5-67ae-488d-a1e7-003d7db4267f\") " pod="openstack/neutron-5fbb8d5d4d-g2v89" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.622670 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9a51d9a5-67ae-488d-a1e7-003d7db4267f-config\") pod \"neutron-5fbb8d5d4d-g2v89\" (UID: \"9a51d9a5-67ae-488d-a1e7-003d7db4267f\") " pod="openstack/neutron-5fbb8d5d4d-g2v89" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.621860 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c9be753-da64-4599-b5d1-baf7f9be58d4-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-zvfqz\" (UID: \"5c9be753-da64-4599-b5d1-baf7f9be58d4\") " pod="openstack/dnsmasq-dns-84b966f6c9-zvfqz" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.622382 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5c9be753-da64-4599-b5d1-baf7f9be58d4-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-zvfqz\" (UID: \"5c9be753-da64-4599-b5d1-baf7f9be58d4\") " pod="openstack/dnsmasq-dns-84b966f6c9-zvfqz" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.622032 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5c9be753-da64-4599-b5d1-baf7f9be58d4-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-zvfqz\" (UID: \"5c9be753-da64-4599-b5d1-baf7f9be58d4\") " pod="openstack/dnsmasq-dns-84b966f6c9-zvfqz" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.622934 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c9be753-da64-4599-b5d1-baf7f9be58d4-config\") pod \"dnsmasq-dns-84b966f6c9-zvfqz\" (UID: \"5c9be753-da64-4599-b5d1-baf7f9be58d4\") " pod="openstack/dnsmasq-dns-84b966f6c9-zvfqz" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.623381 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5c9be753-da64-4599-b5d1-baf7f9be58d4-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-zvfqz\" (UID: \"5c9be753-da64-4599-b5d1-baf7f9be58d4\") " pod="openstack/dnsmasq-dns-84b966f6c9-zvfqz" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.654372 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vt6lb\" (UniqueName: \"kubernetes.io/projected/5c9be753-da64-4599-b5d1-baf7f9be58d4-kube-api-access-vt6lb\") pod \"dnsmasq-dns-84b966f6c9-zvfqz\" (UID: \"5c9be753-da64-4599-b5d1-baf7f9be58d4\") " pod="openstack/dnsmasq-dns-84b966f6c9-zvfqz" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.726574 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a51d9a5-67ae-488d-a1e7-003d7db4267f-ovndb-tls-certs\") pod \"neutron-5fbb8d5d4d-g2v89\" (UID: \"9a51d9a5-67ae-488d-a1e7-003d7db4267f\") " pod="openstack/neutron-5fbb8d5d4d-g2v89" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.726644 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zclvs\" (UniqueName: \"kubernetes.io/projected/9a51d9a5-67ae-488d-a1e7-003d7db4267f-kube-api-access-zclvs\") pod \"neutron-5fbb8d5d4d-g2v89\" (UID: \"9a51d9a5-67ae-488d-a1e7-003d7db4267f\") " pod="openstack/neutron-5fbb8d5d4d-g2v89" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.726721 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a51d9a5-67ae-488d-a1e7-003d7db4267f-combined-ca-bundle\") pod \"neutron-5fbb8d5d4d-g2v89\" (UID: \"9a51d9a5-67ae-488d-a1e7-003d7db4267f\") " pod="openstack/neutron-5fbb8d5d4d-g2v89" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.726785 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9a51d9a5-67ae-488d-a1e7-003d7db4267f-httpd-config\") pod \"neutron-5fbb8d5d4d-g2v89\" (UID: \"9a51d9a5-67ae-488d-a1e7-003d7db4267f\") " pod="openstack/neutron-5fbb8d5d4d-g2v89" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.726808 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9a51d9a5-67ae-488d-a1e7-003d7db4267f-config\") pod \"neutron-5fbb8d5d4d-g2v89\" (UID: \"9a51d9a5-67ae-488d-a1e7-003d7db4267f\") " pod="openstack/neutron-5fbb8d5d4d-g2v89" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.733641 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-94dfcc48d-mlrjd" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.733740 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-94dfcc48d-mlrjd" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.737836 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9a51d9a5-67ae-488d-a1e7-003d7db4267f-httpd-config\") pod \"neutron-5fbb8d5d4d-g2v89\" (UID: \"9a51d9a5-67ae-488d-a1e7-003d7db4267f\") " pod="openstack/neutron-5fbb8d5d4d-g2v89" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.738331 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/9a51d9a5-67ae-488d-a1e7-003d7db4267f-config\") pod \"neutron-5fbb8d5d4d-g2v89\" (UID: \"9a51d9a5-67ae-488d-a1e7-003d7db4267f\") " pod="openstack/neutron-5fbb8d5d4d-g2v89" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.743169 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a51d9a5-67ae-488d-a1e7-003d7db4267f-ovndb-tls-certs\") pod \"neutron-5fbb8d5d4d-g2v89\" (UID: \"9a51d9a5-67ae-488d-a1e7-003d7db4267f\") " pod="openstack/neutron-5fbb8d5d4d-g2v89" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.744349 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zclvs\" (UniqueName: \"kubernetes.io/projected/9a51d9a5-67ae-488d-a1e7-003d7db4267f-kube-api-access-zclvs\") pod \"neutron-5fbb8d5d4d-g2v89\" (UID: \"9a51d9a5-67ae-488d-a1e7-003d7db4267f\") " pod="openstack/neutron-5fbb8d5d4d-g2v89" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.744884 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a51d9a5-67ae-488d-a1e7-003d7db4267f-combined-ca-bundle\") pod \"neutron-5fbb8d5d4d-g2v89\" (UID: \"9a51d9a5-67ae-488d-a1e7-003d7db4267f\") " pod="openstack/neutron-5fbb8d5d4d-g2v89" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.802653 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-zvfqz" Sep 30 14:13:08 crc kubenswrapper[4840]: I0930 14:13:08.927986 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5fbb8d5d4d-g2v89" Sep 30 14:13:09 crc kubenswrapper[4840]: I0930 14:13:09.360282 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"91a0f20e-5a31-42fc-ba4e-c5abd44d497c","Type":"ContainerStarted","Data":"5b72d9d3949a4229e5994d939fa0d0b8dad292749700aa3b7cb583151f686365"} Sep 30 14:13:09 crc kubenswrapper[4840]: I0930 14:13:09.365156 4840 generic.go:334] "Generic (PLEG): container finished" podID="a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa" containerID="801550a13d2831db01a9f0874cc571e7749f7e3f4095df0690b2f7861f87a377" exitCode=0 Sep 30 14:13:09 crc kubenswrapper[4840]: I0930 14:13:09.365270 4840 generic.go:334] "Generic (PLEG): container finished" podID="a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa" containerID="8150ffc2faf4a5d465566a1a3e0e9f2d18a03857ffcd0eba0fdac856a4125a9b" exitCode=143 Sep 30 14:13:09 crc kubenswrapper[4840]: I0930 14:13:09.366729 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa","Type":"ContainerDied","Data":"801550a13d2831db01a9f0874cc571e7749f7e3f4095df0690b2f7861f87a377"} Sep 30 14:13:09 crc kubenswrapper[4840]: I0930 14:13:09.366783 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa","Type":"ContainerDied","Data":"8150ffc2faf4a5d465566a1a3e0e9f2d18a03857ffcd0eba0fdac856a4125a9b"} Sep 30 14:13:09 crc kubenswrapper[4840]: I0930 14:13:09.409704 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 14:13:09 crc kubenswrapper[4840]: I0930 14:13:09.430117 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-zvfqz"] Sep 30 14:13:09 crc kubenswrapper[4840]: I0930 14:13:09.439273 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7sqqs\" (UniqueName: \"kubernetes.io/projected/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-kube-api-access-7sqqs\") pod \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\" (UID: \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\") " Sep 30 14:13:09 crc kubenswrapper[4840]: I0930 14:13:09.439348 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-httpd-run\") pod \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\" (UID: \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\") " Sep 30 14:13:09 crc kubenswrapper[4840]: I0930 14:13:09.439412 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-scripts\") pod \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\" (UID: \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\") " Sep 30 14:13:09 crc kubenswrapper[4840]: I0930 14:13:09.439485 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-logs\") pod \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\" (UID: \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\") " Sep 30 14:13:09 crc kubenswrapper[4840]: I0930 14:13:09.439506 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-internal-tls-certs\") pod \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\" (UID: \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\") " Sep 30 14:13:09 crc kubenswrapper[4840]: I0930 14:13:09.439523 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-combined-ca-bundle\") pod \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\" (UID: \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\") " Sep 30 14:13:09 crc kubenswrapper[4840]: I0930 14:13:09.439591 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-config-data\") pod \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\" (UID: \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\") " Sep 30 14:13:09 crc kubenswrapper[4840]: I0930 14:13:09.439662 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\" (UID: \"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa\") " Sep 30 14:13:09 crc kubenswrapper[4840]: I0930 14:13:09.442013 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa" (UID: "a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:13:09 crc kubenswrapper[4840]: I0930 14:13:09.447915 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-logs" (OuterVolumeSpecName: "logs") pod "a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa" (UID: "a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:13:09 crc kubenswrapper[4840]: I0930 14:13:09.452715 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-scripts" (OuterVolumeSpecName: "scripts") pod "a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa" (UID: "a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:09 crc kubenswrapper[4840]: I0930 14:13:09.456573 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa" (UID: "a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 30 14:13:09 crc kubenswrapper[4840]: I0930 14:13:09.467575 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-kube-api-access-7sqqs" (OuterVolumeSpecName: "kube-api-access-7sqqs") pod "a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa" (UID: "a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa"). InnerVolumeSpecName "kube-api-access-7sqqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:13:09 crc kubenswrapper[4840]: I0930 14:13:09.499772 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa" (UID: "a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:09 crc kubenswrapper[4840]: I0930 14:13:09.504061 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-config-data" (OuterVolumeSpecName: "config-data") pod "a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa" (UID: "a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:09 crc kubenswrapper[4840]: I0930 14:13:09.541545 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:09 crc kubenswrapper[4840]: I0930 14:13:09.541595 4840 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Sep 30 14:13:09 crc kubenswrapper[4840]: I0930 14:13:09.541606 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7sqqs\" (UniqueName: \"kubernetes.io/projected/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-kube-api-access-7sqqs\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:09 crc kubenswrapper[4840]: I0930 14:13:09.541615 4840 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-httpd-run\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:09 crc kubenswrapper[4840]: I0930 14:13:09.541623 4840 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:09 crc kubenswrapper[4840]: I0930 14:13:09.541630 4840 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-logs\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:09 crc kubenswrapper[4840]: I0930 14:13:09.541637 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:09 crc kubenswrapper[4840]: I0930 14:13:09.541726 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa" (UID: "a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:09 crc kubenswrapper[4840]: I0930 14:13:09.558136 4840 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Sep 30 14:13:09 crc kubenswrapper[4840]: I0930 14:13:09.644490 4840 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:09 crc kubenswrapper[4840]: I0930 14:13:09.644519 4840 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:09 crc kubenswrapper[4840]: I0930 14:13:09.672992 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5cc547b489-2p58b" Sep 30 14:13:09 crc kubenswrapper[4840]: I0930 14:13:09.723559 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5fbb8d5d4d-g2v89"] Sep 30 14:13:09 crc kubenswrapper[4840]: I0930 14:13:09.836259 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7f785c99fc-fpf4g" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.388744 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa","Type":"ContainerDied","Data":"ed8ad838a3d1aa034d4a67b2b948166a2bb568d0ef615fbdd6094de64d36d483"} Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.389013 4840 scope.go:117] "RemoveContainer" containerID="801550a13d2831db01a9f0874cc571e7749f7e3f4095df0690b2f7861f87a377" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.389120 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.401598 4840 generic.go:334] "Generic (PLEG): container finished" podID="5c9be753-da64-4599-b5d1-baf7f9be58d4" containerID="b59efce57c2c6f84cdf6482bb6869209f49de47b514336e597e1121d41db727d" exitCode=0 Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.401654 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-zvfqz" event={"ID":"5c9be753-da64-4599-b5d1-baf7f9be58d4","Type":"ContainerDied","Data":"b59efce57c2c6f84cdf6482bb6869209f49de47b514336e597e1121d41db727d"} Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.401677 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-zvfqz" event={"ID":"5c9be753-da64-4599-b5d1-baf7f9be58d4","Type":"ContainerStarted","Data":"e5428d36f647cdf3937529312de47d20ee328949c756a09885e9431a5246bb6c"} Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.415384 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.428705 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.430645 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5fbb8d5d4d-g2v89" event={"ID":"9a51d9a5-67ae-488d-a1e7-003d7db4267f","Type":"ContainerStarted","Data":"889bc6c8cd3512a46aa4cd11e1a67d3387b504acec3381ff67d135c78f8e8ccb"} Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.430688 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5fbb8d5d4d-g2v89" event={"ID":"9a51d9a5-67ae-488d-a1e7-003d7db4267f","Type":"ContainerStarted","Data":"058945b169bf7708ddcedb1ca2e22711684ab3b132c19cf217a8cab6341ef1d9"} Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.430768 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="91a0f20e-5a31-42fc-ba4e-c5abd44d497c" containerName="glance-log" containerID="cri-o://cdb11879b0067fe94fa10997b5f54c42a9263b63162c6e54eb8ee403ec65cd57" gracePeriod=30 Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.431045 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="91a0f20e-5a31-42fc-ba4e-c5abd44d497c" containerName="glance-httpd" containerID="cri-o://5b72d9d3949a4229e5994d939fa0d0b8dad292749700aa3b7cb583151f686365" gracePeriod=30 Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.450596 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:13:10 crc kubenswrapper[4840]: E0930 14:13:10.451038 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa" containerName="glance-httpd" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.451063 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa" containerName="glance-httpd" Sep 30 14:13:10 crc kubenswrapper[4840]: E0930 14:13:10.451093 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa" containerName="glance-log" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.451099 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa" containerName="glance-log" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.451327 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa" containerName="glance-log" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.451338 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa" containerName="glance-httpd" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.458920 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.464022 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.466123 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.488674 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.492821 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=25.492801342 podStartE2EDuration="25.492801342s" podCreationTimestamp="2025-09-30 14:12:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:13:10.475045909 +0000 UTC m=+1019.104132332" watchObservedRunningTime="2025-09-30 14:13:10.492801342 +0000 UTC m=+1019.121887765" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.562751 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c499ea9a-7047-493b-a529-1b5115119b25-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c499ea9a-7047-493b-a529-1b5115119b25\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.562822 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"c499ea9a-7047-493b-a529-1b5115119b25\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.562854 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c499ea9a-7047-493b-a529-1b5115119b25-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c499ea9a-7047-493b-a529-1b5115119b25\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.562881 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c499ea9a-7047-493b-a529-1b5115119b25-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c499ea9a-7047-493b-a529-1b5115119b25\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.562914 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c499ea9a-7047-493b-a529-1b5115119b25-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c499ea9a-7047-493b-a529-1b5115119b25\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.562950 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hs4t6\" (UniqueName: \"kubernetes.io/projected/c499ea9a-7047-493b-a529-1b5115119b25-kube-api-access-hs4t6\") pod \"glance-default-internal-api-0\" (UID: \"c499ea9a-7047-493b-a529-1b5115119b25\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.563001 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c499ea9a-7047-493b-a529-1b5115119b25-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c499ea9a-7047-493b-a529-1b5115119b25\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.563036 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c499ea9a-7047-493b-a529-1b5115119b25-logs\") pod \"glance-default-internal-api-0\" (UID: \"c499ea9a-7047-493b-a529-1b5115119b25\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.665229 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c499ea9a-7047-493b-a529-1b5115119b25-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c499ea9a-7047-493b-a529-1b5115119b25\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.665280 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"c499ea9a-7047-493b-a529-1b5115119b25\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.665303 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c499ea9a-7047-493b-a529-1b5115119b25-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c499ea9a-7047-493b-a529-1b5115119b25\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.665328 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c499ea9a-7047-493b-a529-1b5115119b25-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c499ea9a-7047-493b-a529-1b5115119b25\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.665355 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c499ea9a-7047-493b-a529-1b5115119b25-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c499ea9a-7047-493b-a529-1b5115119b25\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.665386 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hs4t6\" (UniqueName: \"kubernetes.io/projected/c499ea9a-7047-493b-a529-1b5115119b25-kube-api-access-hs4t6\") pod \"glance-default-internal-api-0\" (UID: \"c499ea9a-7047-493b-a529-1b5115119b25\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.665433 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c499ea9a-7047-493b-a529-1b5115119b25-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c499ea9a-7047-493b-a529-1b5115119b25\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.665510 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c499ea9a-7047-493b-a529-1b5115119b25-logs\") pod \"glance-default-internal-api-0\" (UID: \"c499ea9a-7047-493b-a529-1b5115119b25\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.665942 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c499ea9a-7047-493b-a529-1b5115119b25-logs\") pod \"glance-default-internal-api-0\" (UID: \"c499ea9a-7047-493b-a529-1b5115119b25\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.666329 4840 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"c499ea9a-7047-493b-a529-1b5115119b25\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-internal-api-0" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.672318 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c499ea9a-7047-493b-a529-1b5115119b25-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c499ea9a-7047-493b-a529-1b5115119b25\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.674344 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c499ea9a-7047-493b-a529-1b5115119b25-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c499ea9a-7047-493b-a529-1b5115119b25\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.674365 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c499ea9a-7047-493b-a529-1b5115119b25-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c499ea9a-7047-493b-a529-1b5115119b25\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.681291 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c499ea9a-7047-493b-a529-1b5115119b25-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c499ea9a-7047-493b-a529-1b5115119b25\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.690517 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c499ea9a-7047-493b-a529-1b5115119b25-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c499ea9a-7047-493b-a529-1b5115119b25\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.696207 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hs4t6\" (UniqueName: \"kubernetes.io/projected/c499ea9a-7047-493b-a529-1b5115119b25-kube-api-access-hs4t6\") pod \"glance-default-internal-api-0\" (UID: \"c499ea9a-7047-493b-a529-1b5115119b25\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.722235 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"c499ea9a-7047-493b-a529-1b5115119b25\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.782207 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.833860 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-658bb978b7-lrh6p"] Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.835495 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-658bb978b7-lrh6p" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.847108 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.847376 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Sep 30 14:13:10 crc kubenswrapper[4840]: I0930 14:13:10.865388 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-658bb978b7-lrh6p"] Sep 30 14:13:11 crc kubenswrapper[4840]: I0930 14:13:11.010003 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8e55cd0-dac0-402d-beb7-6482894554fb-combined-ca-bundle\") pod \"neutron-658bb978b7-lrh6p\" (UID: \"e8e55cd0-dac0-402d-beb7-6482894554fb\") " pod="openstack/neutron-658bb978b7-lrh6p" Sep 30 14:13:11 crc kubenswrapper[4840]: I0930 14:13:11.010099 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8e55cd0-dac0-402d-beb7-6482894554fb-public-tls-certs\") pod \"neutron-658bb978b7-lrh6p\" (UID: \"e8e55cd0-dac0-402d-beb7-6482894554fb\") " pod="openstack/neutron-658bb978b7-lrh6p" Sep 30 14:13:11 crc kubenswrapper[4840]: I0930 14:13:11.010125 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8e55cd0-dac0-402d-beb7-6482894554fb-internal-tls-certs\") pod \"neutron-658bb978b7-lrh6p\" (UID: \"e8e55cd0-dac0-402d-beb7-6482894554fb\") " pod="openstack/neutron-658bb978b7-lrh6p" Sep 30 14:13:11 crc kubenswrapper[4840]: I0930 14:13:11.010162 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e8e55cd0-dac0-402d-beb7-6482894554fb-httpd-config\") pod \"neutron-658bb978b7-lrh6p\" (UID: \"e8e55cd0-dac0-402d-beb7-6482894554fb\") " pod="openstack/neutron-658bb978b7-lrh6p" Sep 30 14:13:11 crc kubenswrapper[4840]: I0930 14:13:11.010209 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8e55cd0-dac0-402d-beb7-6482894554fb-ovndb-tls-certs\") pod \"neutron-658bb978b7-lrh6p\" (UID: \"e8e55cd0-dac0-402d-beb7-6482894554fb\") " pod="openstack/neutron-658bb978b7-lrh6p" Sep 30 14:13:11 crc kubenswrapper[4840]: I0930 14:13:11.010239 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e8e55cd0-dac0-402d-beb7-6482894554fb-config\") pod \"neutron-658bb978b7-lrh6p\" (UID: \"e8e55cd0-dac0-402d-beb7-6482894554fb\") " pod="openstack/neutron-658bb978b7-lrh6p" Sep 30 14:13:11 crc kubenswrapper[4840]: I0930 14:13:11.010280 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7j5q\" (UniqueName: \"kubernetes.io/projected/e8e55cd0-dac0-402d-beb7-6482894554fb-kube-api-access-b7j5q\") pod \"neutron-658bb978b7-lrh6p\" (UID: \"e8e55cd0-dac0-402d-beb7-6482894554fb\") " pod="openstack/neutron-658bb978b7-lrh6p" Sep 30 14:13:11 crc kubenswrapper[4840]: I0930 14:13:11.111742 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e8e55cd0-dac0-402d-beb7-6482894554fb-config\") pod \"neutron-658bb978b7-lrh6p\" (UID: \"e8e55cd0-dac0-402d-beb7-6482894554fb\") " pod="openstack/neutron-658bb978b7-lrh6p" Sep 30 14:13:11 crc kubenswrapper[4840]: I0930 14:13:11.111840 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7j5q\" (UniqueName: \"kubernetes.io/projected/e8e55cd0-dac0-402d-beb7-6482894554fb-kube-api-access-b7j5q\") pod \"neutron-658bb978b7-lrh6p\" (UID: \"e8e55cd0-dac0-402d-beb7-6482894554fb\") " pod="openstack/neutron-658bb978b7-lrh6p" Sep 30 14:13:11 crc kubenswrapper[4840]: I0930 14:13:11.111908 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8e55cd0-dac0-402d-beb7-6482894554fb-combined-ca-bundle\") pod \"neutron-658bb978b7-lrh6p\" (UID: \"e8e55cd0-dac0-402d-beb7-6482894554fb\") " pod="openstack/neutron-658bb978b7-lrh6p" Sep 30 14:13:11 crc kubenswrapper[4840]: I0930 14:13:11.112003 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8e55cd0-dac0-402d-beb7-6482894554fb-public-tls-certs\") pod \"neutron-658bb978b7-lrh6p\" (UID: \"e8e55cd0-dac0-402d-beb7-6482894554fb\") " pod="openstack/neutron-658bb978b7-lrh6p" Sep 30 14:13:11 crc kubenswrapper[4840]: I0930 14:13:11.112024 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8e55cd0-dac0-402d-beb7-6482894554fb-internal-tls-certs\") pod \"neutron-658bb978b7-lrh6p\" (UID: \"e8e55cd0-dac0-402d-beb7-6482894554fb\") " pod="openstack/neutron-658bb978b7-lrh6p" Sep 30 14:13:11 crc kubenswrapper[4840]: I0930 14:13:11.112050 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e8e55cd0-dac0-402d-beb7-6482894554fb-httpd-config\") pod \"neutron-658bb978b7-lrh6p\" (UID: \"e8e55cd0-dac0-402d-beb7-6482894554fb\") " pod="openstack/neutron-658bb978b7-lrh6p" Sep 30 14:13:11 crc kubenswrapper[4840]: I0930 14:13:11.112094 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8e55cd0-dac0-402d-beb7-6482894554fb-ovndb-tls-certs\") pod \"neutron-658bb978b7-lrh6p\" (UID: \"e8e55cd0-dac0-402d-beb7-6482894554fb\") " pod="openstack/neutron-658bb978b7-lrh6p" Sep 30 14:13:11 crc kubenswrapper[4840]: I0930 14:13:11.121569 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e8e55cd0-dac0-402d-beb7-6482894554fb-config\") pod \"neutron-658bb978b7-lrh6p\" (UID: \"e8e55cd0-dac0-402d-beb7-6482894554fb\") " pod="openstack/neutron-658bb978b7-lrh6p" Sep 30 14:13:11 crc kubenswrapper[4840]: I0930 14:13:11.122717 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8e55cd0-dac0-402d-beb7-6482894554fb-public-tls-certs\") pod \"neutron-658bb978b7-lrh6p\" (UID: \"e8e55cd0-dac0-402d-beb7-6482894554fb\") " pod="openstack/neutron-658bb978b7-lrh6p" Sep 30 14:13:11 crc kubenswrapper[4840]: I0930 14:13:11.125300 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8e55cd0-dac0-402d-beb7-6482894554fb-internal-tls-certs\") pod \"neutron-658bb978b7-lrh6p\" (UID: \"e8e55cd0-dac0-402d-beb7-6482894554fb\") " pod="openstack/neutron-658bb978b7-lrh6p" Sep 30 14:13:11 crc kubenswrapper[4840]: I0930 14:13:11.126280 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8e55cd0-dac0-402d-beb7-6482894554fb-ovndb-tls-certs\") pod \"neutron-658bb978b7-lrh6p\" (UID: \"e8e55cd0-dac0-402d-beb7-6482894554fb\") " pod="openstack/neutron-658bb978b7-lrh6p" Sep 30 14:13:11 crc kubenswrapper[4840]: I0930 14:13:11.126825 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8e55cd0-dac0-402d-beb7-6482894554fb-combined-ca-bundle\") pod \"neutron-658bb978b7-lrh6p\" (UID: \"e8e55cd0-dac0-402d-beb7-6482894554fb\") " pod="openstack/neutron-658bb978b7-lrh6p" Sep 30 14:13:11 crc kubenswrapper[4840]: I0930 14:13:11.132312 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e8e55cd0-dac0-402d-beb7-6482894554fb-httpd-config\") pod \"neutron-658bb978b7-lrh6p\" (UID: \"e8e55cd0-dac0-402d-beb7-6482894554fb\") " pod="openstack/neutron-658bb978b7-lrh6p" Sep 30 14:13:11 crc kubenswrapper[4840]: I0930 14:13:11.151793 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7j5q\" (UniqueName: \"kubernetes.io/projected/e8e55cd0-dac0-402d-beb7-6482894554fb-kube-api-access-b7j5q\") pod \"neutron-658bb978b7-lrh6p\" (UID: \"e8e55cd0-dac0-402d-beb7-6482894554fb\") " pod="openstack/neutron-658bb978b7-lrh6p" Sep 30 14:13:11 crc kubenswrapper[4840]: I0930 14:13:11.176139 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-658bb978b7-lrh6p" Sep 30 14:13:11 crc kubenswrapper[4840]: I0930 14:13:11.451758 4840 generic.go:334] "Generic (PLEG): container finished" podID="91a0f20e-5a31-42fc-ba4e-c5abd44d497c" containerID="5b72d9d3949a4229e5994d939fa0d0b8dad292749700aa3b7cb583151f686365" exitCode=0 Sep 30 14:13:11 crc kubenswrapper[4840]: I0930 14:13:11.451796 4840 generic.go:334] "Generic (PLEG): container finished" podID="91a0f20e-5a31-42fc-ba4e-c5abd44d497c" containerID="cdb11879b0067fe94fa10997b5f54c42a9263b63162c6e54eb8ee403ec65cd57" exitCode=143 Sep 30 14:13:11 crc kubenswrapper[4840]: I0930 14:13:11.451826 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"91a0f20e-5a31-42fc-ba4e-c5abd44d497c","Type":"ContainerDied","Data":"5b72d9d3949a4229e5994d939fa0d0b8dad292749700aa3b7cb583151f686365"} Sep 30 14:13:11 crc kubenswrapper[4840]: I0930 14:13:11.451878 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"91a0f20e-5a31-42fc-ba4e-c5abd44d497c","Type":"ContainerDied","Data":"cdb11879b0067fe94fa10997b5f54c42a9263b63162c6e54eb8ee403ec65cd57"} Sep 30 14:13:11 crc kubenswrapper[4840]: I0930 14:13:11.454147 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5fbb8d5d4d-g2v89" event={"ID":"9a51d9a5-67ae-488d-a1e7-003d7db4267f","Type":"ContainerStarted","Data":"a1c3a2c622f9191ac9262da20051252eb62e825475ca9959fec3292d9d52112d"} Sep 30 14:13:11 crc kubenswrapper[4840]: I0930 14:13:11.454296 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5fbb8d5d4d-g2v89" Sep 30 14:13:11 crc kubenswrapper[4840]: I0930 14:13:11.478991 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5fbb8d5d4d-g2v89" podStartSLOduration=3.478966799 podStartE2EDuration="3.478966799s" podCreationTimestamp="2025-09-30 14:13:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:13:11.473738163 +0000 UTC m=+1020.102824606" watchObservedRunningTime="2025-09-30 14:13:11.478966799 +0000 UTC m=+1020.108053222" Sep 30 14:13:12 crc kubenswrapper[4840]: I0930 14:13:12.134503 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa" path="/var/lib/kubelet/pods/a7b2bbcc-8c5b-4eaa-ab08-87a738b4adaa/volumes" Sep 30 14:13:12 crc kubenswrapper[4840]: I0930 14:13:12.464393 4840 generic.go:334] "Generic (PLEG): container finished" podID="0f907c87-9649-4dd2-a419-97055feb4d18" containerID="25f1bac61cdc7bb81eac30ec24cc17df033b36b1740140f97751b58d10a731b7" exitCode=0 Sep 30 14:13:12 crc kubenswrapper[4840]: I0930 14:13:12.464458 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5gzc5" event={"ID":"0f907c87-9649-4dd2-a419-97055feb4d18","Type":"ContainerDied","Data":"25f1bac61cdc7bb81eac30ec24cc17df033b36b1740140f97751b58d10a731b7"} Sep 30 14:13:12 crc kubenswrapper[4840]: I0930 14:13:12.469051 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-788d88bd6f-l44pw" Sep 30 14:13:14 crc kubenswrapper[4840]: I0930 14:13:14.744660 4840 scope.go:117] "RemoveContainer" containerID="8150ffc2faf4a5d465566a1a3e0e9f2d18a03857ffcd0eba0fdac856a4125a9b" Sep 30 14:13:14 crc kubenswrapper[4840]: I0930 14:13:14.906658 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5gzc5" Sep 30 14:13:14 crc kubenswrapper[4840]: I0930 14:13:14.985725 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0f907c87-9649-4dd2-a419-97055feb4d18-credential-keys\") pod \"0f907c87-9649-4dd2-a419-97055feb4d18\" (UID: \"0f907c87-9649-4dd2-a419-97055feb4d18\") " Sep 30 14:13:14 crc kubenswrapper[4840]: I0930 14:13:14.985760 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f907c87-9649-4dd2-a419-97055feb4d18-config-data\") pod \"0f907c87-9649-4dd2-a419-97055feb4d18\" (UID: \"0f907c87-9649-4dd2-a419-97055feb4d18\") " Sep 30 14:13:14 crc kubenswrapper[4840]: I0930 14:13:14.985789 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f907c87-9649-4dd2-a419-97055feb4d18-scripts\") pod \"0f907c87-9649-4dd2-a419-97055feb4d18\" (UID: \"0f907c87-9649-4dd2-a419-97055feb4d18\") " Sep 30 14:13:14 crc kubenswrapper[4840]: I0930 14:13:14.985823 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0f907c87-9649-4dd2-a419-97055feb4d18-fernet-keys\") pod \"0f907c87-9649-4dd2-a419-97055feb4d18\" (UID: \"0f907c87-9649-4dd2-a419-97055feb4d18\") " Sep 30 14:13:14 crc kubenswrapper[4840]: I0930 14:13:14.985843 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59r2n\" (UniqueName: \"kubernetes.io/projected/0f907c87-9649-4dd2-a419-97055feb4d18-kube-api-access-59r2n\") pod \"0f907c87-9649-4dd2-a419-97055feb4d18\" (UID: \"0f907c87-9649-4dd2-a419-97055feb4d18\") " Sep 30 14:13:14 crc kubenswrapper[4840]: I0930 14:13:14.985897 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f907c87-9649-4dd2-a419-97055feb4d18-combined-ca-bundle\") pod \"0f907c87-9649-4dd2-a419-97055feb4d18\" (UID: \"0f907c87-9649-4dd2-a419-97055feb4d18\") " Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.003142 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f907c87-9649-4dd2-a419-97055feb4d18-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "0f907c87-9649-4dd2-a419-97055feb4d18" (UID: "0f907c87-9649-4dd2-a419-97055feb4d18"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.005132 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f907c87-9649-4dd2-a419-97055feb4d18-kube-api-access-59r2n" (OuterVolumeSpecName: "kube-api-access-59r2n") pod "0f907c87-9649-4dd2-a419-97055feb4d18" (UID: "0f907c87-9649-4dd2-a419-97055feb4d18"). InnerVolumeSpecName "kube-api-access-59r2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.005359 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f907c87-9649-4dd2-a419-97055feb4d18-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "0f907c87-9649-4dd2-a419-97055feb4d18" (UID: "0f907c87-9649-4dd2-a419-97055feb4d18"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.013284 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f907c87-9649-4dd2-a419-97055feb4d18-scripts" (OuterVolumeSpecName: "scripts") pod "0f907c87-9649-4dd2-a419-97055feb4d18" (UID: "0f907c87-9649-4dd2-a419-97055feb4d18"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.045780 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f907c87-9649-4dd2-a419-97055feb4d18-config-data" (OuterVolumeSpecName: "config-data") pod "0f907c87-9649-4dd2-a419-97055feb4d18" (UID: "0f907c87-9649-4dd2-a419-97055feb4d18"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.059756 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f907c87-9649-4dd2-a419-97055feb4d18-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0f907c87-9649-4dd2-a419-97055feb4d18" (UID: "0f907c87-9649-4dd2-a419-97055feb4d18"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.087053 4840 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0f907c87-9649-4dd2-a419-97055feb4d18-credential-keys\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.087376 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f907c87-9649-4dd2-a419-97055feb4d18-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.087388 4840 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f907c87-9649-4dd2-a419-97055feb4d18-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.087396 4840 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0f907c87-9649-4dd2-a419-97055feb4d18-fernet-keys\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.087406 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59r2n\" (UniqueName: \"kubernetes.io/projected/0f907c87-9649-4dd2-a419-97055feb4d18-kube-api-access-59r2n\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.087418 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f907c87-9649-4dd2-a419-97055feb4d18-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.311595 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.391599 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-658bb978b7-lrh6p"] Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.393051 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-public-tls-certs\") pod \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\" (UID: \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\") " Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.393164 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-logs\") pod \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\" (UID: \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\") " Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.393204 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-scripts\") pod \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\" (UID: \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\") " Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.393245 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wb5df\" (UniqueName: \"kubernetes.io/projected/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-kube-api-access-wb5df\") pod \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\" (UID: \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\") " Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.393284 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-httpd-run\") pod \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\" (UID: \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\") " Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.393346 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\" (UID: \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\") " Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.393411 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-combined-ca-bundle\") pod \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\" (UID: \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\") " Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.393443 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-config-data\") pod \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\" (UID: \"91a0f20e-5a31-42fc-ba4e-c5abd44d497c\") " Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.394575 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "91a0f20e-5a31-42fc-ba4e-c5abd44d497c" (UID: "91a0f20e-5a31-42fc-ba4e-c5abd44d497c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.394623 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-logs" (OuterVolumeSpecName: "logs") pod "91a0f20e-5a31-42fc-ba4e-c5abd44d497c" (UID: "91a0f20e-5a31-42fc-ba4e-c5abd44d497c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.398468 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-scripts" (OuterVolumeSpecName: "scripts") pod "91a0f20e-5a31-42fc-ba4e-c5abd44d497c" (UID: "91a0f20e-5a31-42fc-ba4e-c5abd44d497c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.399433 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "91a0f20e-5a31-42fc-ba4e-c5abd44d497c" (UID: "91a0f20e-5a31-42fc-ba4e-c5abd44d497c"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.400763 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-kube-api-access-wb5df" (OuterVolumeSpecName: "kube-api-access-wb5df") pod "91a0f20e-5a31-42fc-ba4e-c5abd44d497c" (UID: "91a0f20e-5a31-42fc-ba4e-c5abd44d497c"). InnerVolumeSpecName "kube-api-access-wb5df". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.427075 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "91a0f20e-5a31-42fc-ba4e-c5abd44d497c" (UID: "91a0f20e-5a31-42fc-ba4e-c5abd44d497c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.450507 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-config-data" (OuterVolumeSpecName: "config-data") pod "91a0f20e-5a31-42fc-ba4e-c5abd44d497c" (UID: "91a0f20e-5a31-42fc-ba4e-c5abd44d497c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.452646 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "91a0f20e-5a31-42fc-ba4e-c5abd44d497c" (UID: "91a0f20e-5a31-42fc-ba4e-c5abd44d497c"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.489456 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.490814 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"91a0f20e-5a31-42fc-ba4e-c5abd44d497c","Type":"ContainerDied","Data":"669f521351fd5967336af154cee4ccd531f6c3667cecc3c32708eb943d74318f"} Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.490906 4840 scope.go:117] "RemoveContainer" containerID="5b72d9d3949a4229e5994d939fa0d0b8dad292749700aa3b7cb583151f686365" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.494829 4840 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-logs\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.494853 4840 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.494863 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wb5df\" (UniqueName: \"kubernetes.io/projected/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-kube-api-access-wb5df\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.494871 4840 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-httpd-run\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.494895 4840 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.494905 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.494913 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.494920 4840 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/91a0f20e-5a31-42fc-ba4e-c5abd44d497c-public-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.495139 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-658bb978b7-lrh6p" event={"ID":"e8e55cd0-dac0-402d-beb7-6482894554fb","Type":"ContainerStarted","Data":"12ad051f05c193e615b64f8a9115b3fdaa299e8f53c2ee90c3c2281a48a47181"} Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.500327 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5gzc5" event={"ID":"0f907c87-9649-4dd2-a419-97055feb4d18","Type":"ContainerDied","Data":"f3c9b387244c00a461e6391af9513520d1ff8adb2a889327390ea106c5a648b0"} Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.500362 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3c9b387244c00a461e6391af9513520d1ff8adb2a889327390ea106c5a648b0" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.500417 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5gzc5" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.514027 4840 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.535681 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.539659 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.549977 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.559199 4840 scope.go:117] "RemoveContainer" containerID="cdb11879b0067fe94fa10997b5f54c42a9263b63162c6e54eb8ee403ec65cd57" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.565955 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 14:13:15 crc kubenswrapper[4840]: E0930 14:13:15.566508 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f907c87-9649-4dd2-a419-97055feb4d18" containerName="keystone-bootstrap" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.566626 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f907c87-9649-4dd2-a419-97055feb4d18" containerName="keystone-bootstrap" Sep 30 14:13:15 crc kubenswrapper[4840]: E0930 14:13:15.566723 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91a0f20e-5a31-42fc-ba4e-c5abd44d497c" containerName="glance-httpd" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.566793 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="91a0f20e-5a31-42fc-ba4e-c5abd44d497c" containerName="glance-httpd" Sep 30 14:13:15 crc kubenswrapper[4840]: E0930 14:13:15.566891 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91a0f20e-5a31-42fc-ba4e-c5abd44d497c" containerName="glance-log" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.566944 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="91a0f20e-5a31-42fc-ba4e-c5abd44d497c" containerName="glance-log" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.567166 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="91a0f20e-5a31-42fc-ba4e-c5abd44d497c" containerName="glance-httpd" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.567253 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="91a0f20e-5a31-42fc-ba4e-c5abd44d497c" containerName="glance-log" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.567311 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f907c87-9649-4dd2-a419-97055feb4d18" containerName="keystone-bootstrap" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.576056 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.592924 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.607856 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.611160 4840 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.617624 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.713205 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72l8s\" (UniqueName: \"kubernetes.io/projected/b96c744c-10e9-4c0c-9ece-069c0c90e67a-kube-api-access-72l8s\") pod \"glance-default-external-api-0\" (UID: \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\") " pod="openstack/glance-default-external-api-0" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.713269 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b96c744c-10e9-4c0c-9ece-069c0c90e67a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\") " pod="openstack/glance-default-external-api-0" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.713301 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b96c744c-10e9-4c0c-9ece-069c0c90e67a-scripts\") pod \"glance-default-external-api-0\" (UID: \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\") " pod="openstack/glance-default-external-api-0" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.713320 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\") " pod="openstack/glance-default-external-api-0" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.713384 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b96c744c-10e9-4c0c-9ece-069c0c90e67a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\") " pod="openstack/glance-default-external-api-0" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.713405 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b96c744c-10e9-4c0c-9ece-069c0c90e67a-config-data\") pod \"glance-default-external-api-0\" (UID: \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\") " pod="openstack/glance-default-external-api-0" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.713431 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b96c744c-10e9-4c0c-9ece-069c0c90e67a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\") " pod="openstack/glance-default-external-api-0" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.713447 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b96c744c-10e9-4c0c-9ece-069c0c90e67a-logs\") pod \"glance-default-external-api-0\" (UID: \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\") " pod="openstack/glance-default-external-api-0" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.817653 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b96c744c-10e9-4c0c-9ece-069c0c90e67a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\") " pod="openstack/glance-default-external-api-0" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.817707 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b96c744c-10e9-4c0c-9ece-069c0c90e67a-config-data\") pod \"glance-default-external-api-0\" (UID: \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\") " pod="openstack/glance-default-external-api-0" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.817741 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b96c744c-10e9-4c0c-9ece-069c0c90e67a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\") " pod="openstack/glance-default-external-api-0" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.817762 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b96c744c-10e9-4c0c-9ece-069c0c90e67a-logs\") pod \"glance-default-external-api-0\" (UID: \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\") " pod="openstack/glance-default-external-api-0" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.817815 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72l8s\" (UniqueName: \"kubernetes.io/projected/b96c744c-10e9-4c0c-9ece-069c0c90e67a-kube-api-access-72l8s\") pod \"glance-default-external-api-0\" (UID: \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\") " pod="openstack/glance-default-external-api-0" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.817861 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b96c744c-10e9-4c0c-9ece-069c0c90e67a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\") " pod="openstack/glance-default-external-api-0" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.817894 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b96c744c-10e9-4c0c-9ece-069c0c90e67a-scripts\") pod \"glance-default-external-api-0\" (UID: \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\") " pod="openstack/glance-default-external-api-0" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.818193 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\") " pod="openstack/glance-default-external-api-0" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.818558 4840 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.819432 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b96c744c-10e9-4c0c-9ece-069c0c90e67a-logs\") pod \"glance-default-external-api-0\" (UID: \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\") " pod="openstack/glance-default-external-api-0" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.820587 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b96c744c-10e9-4c0c-9ece-069c0c90e67a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\") " pod="openstack/glance-default-external-api-0" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.827274 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b96c744c-10e9-4c0c-9ece-069c0c90e67a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\") " pod="openstack/glance-default-external-api-0" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.830079 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b96c744c-10e9-4c0c-9ece-069c0c90e67a-config-data\") pod \"glance-default-external-api-0\" (UID: \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\") " pod="openstack/glance-default-external-api-0" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.830639 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b96c744c-10e9-4c0c-9ece-069c0c90e67a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\") " pod="openstack/glance-default-external-api-0" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.840385 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72l8s\" (UniqueName: \"kubernetes.io/projected/b96c744c-10e9-4c0c-9ece-069c0c90e67a-kube-api-access-72l8s\") pod \"glance-default-external-api-0\" (UID: \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\") " pod="openstack/glance-default-external-api-0" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.849118 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b96c744c-10e9-4c0c-9ece-069c0c90e67a-scripts\") pod \"glance-default-external-api-0\" (UID: \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\") " pod="openstack/glance-default-external-api-0" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.868109 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\") " pod="openstack/glance-default-external-api-0" Sep 30 14:13:15 crc kubenswrapper[4840]: I0930 14:13:15.934236 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.009979 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-6c78d49dc8-kqrxd"] Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.011061 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6c78d49dc8-kqrxd" Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.022358 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gtvd7" Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.022674 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.022892 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.023597 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.023740 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.023859 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.027417 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6c78d49dc8-kqrxd"] Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.126930 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/692fd2c9-5168-48fb-9909-7b1cab678b75-public-tls-certs\") pod \"keystone-6c78d49dc8-kqrxd\" (UID: \"692fd2c9-5168-48fb-9909-7b1cab678b75\") " pod="openstack/keystone-6c78d49dc8-kqrxd" Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.127317 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/692fd2c9-5168-48fb-9909-7b1cab678b75-fernet-keys\") pod \"keystone-6c78d49dc8-kqrxd\" (UID: \"692fd2c9-5168-48fb-9909-7b1cab678b75\") " pod="openstack/keystone-6c78d49dc8-kqrxd" Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.127348 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/692fd2c9-5168-48fb-9909-7b1cab678b75-combined-ca-bundle\") pod \"keystone-6c78d49dc8-kqrxd\" (UID: \"692fd2c9-5168-48fb-9909-7b1cab678b75\") " pod="openstack/keystone-6c78d49dc8-kqrxd" Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.127417 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/692fd2c9-5168-48fb-9909-7b1cab678b75-internal-tls-certs\") pod \"keystone-6c78d49dc8-kqrxd\" (UID: \"692fd2c9-5168-48fb-9909-7b1cab678b75\") " pod="openstack/keystone-6c78d49dc8-kqrxd" Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.127449 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/692fd2c9-5168-48fb-9909-7b1cab678b75-config-data\") pod \"keystone-6c78d49dc8-kqrxd\" (UID: \"692fd2c9-5168-48fb-9909-7b1cab678b75\") " pod="openstack/keystone-6c78d49dc8-kqrxd" Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.127490 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8l6km\" (UniqueName: \"kubernetes.io/projected/692fd2c9-5168-48fb-9909-7b1cab678b75-kube-api-access-8l6km\") pod \"keystone-6c78d49dc8-kqrxd\" (UID: \"692fd2c9-5168-48fb-9909-7b1cab678b75\") " pod="openstack/keystone-6c78d49dc8-kqrxd" Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.127524 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/692fd2c9-5168-48fb-9909-7b1cab678b75-credential-keys\") pod \"keystone-6c78d49dc8-kqrxd\" (UID: \"692fd2c9-5168-48fb-9909-7b1cab678b75\") " pod="openstack/keystone-6c78d49dc8-kqrxd" Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.127559 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/692fd2c9-5168-48fb-9909-7b1cab678b75-scripts\") pod \"keystone-6c78d49dc8-kqrxd\" (UID: \"692fd2c9-5168-48fb-9909-7b1cab678b75\") " pod="openstack/keystone-6c78d49dc8-kqrxd" Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.138594 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91a0f20e-5a31-42fc-ba4e-c5abd44d497c" path="/var/lib/kubelet/pods/91a0f20e-5a31-42fc-ba4e-c5abd44d497c/volumes" Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.228699 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/692fd2c9-5168-48fb-9909-7b1cab678b75-config-data\") pod \"keystone-6c78d49dc8-kqrxd\" (UID: \"692fd2c9-5168-48fb-9909-7b1cab678b75\") " pod="openstack/keystone-6c78d49dc8-kqrxd" Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.228766 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8l6km\" (UniqueName: \"kubernetes.io/projected/692fd2c9-5168-48fb-9909-7b1cab678b75-kube-api-access-8l6km\") pod \"keystone-6c78d49dc8-kqrxd\" (UID: \"692fd2c9-5168-48fb-9909-7b1cab678b75\") " pod="openstack/keystone-6c78d49dc8-kqrxd" Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.228827 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/692fd2c9-5168-48fb-9909-7b1cab678b75-credential-keys\") pod \"keystone-6c78d49dc8-kqrxd\" (UID: \"692fd2c9-5168-48fb-9909-7b1cab678b75\") " pod="openstack/keystone-6c78d49dc8-kqrxd" Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.228865 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/692fd2c9-5168-48fb-9909-7b1cab678b75-scripts\") pod \"keystone-6c78d49dc8-kqrxd\" (UID: \"692fd2c9-5168-48fb-9909-7b1cab678b75\") " pod="openstack/keystone-6c78d49dc8-kqrxd" Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.228905 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/692fd2c9-5168-48fb-9909-7b1cab678b75-public-tls-certs\") pod \"keystone-6c78d49dc8-kqrxd\" (UID: \"692fd2c9-5168-48fb-9909-7b1cab678b75\") " pod="openstack/keystone-6c78d49dc8-kqrxd" Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.228944 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/692fd2c9-5168-48fb-9909-7b1cab678b75-fernet-keys\") pod \"keystone-6c78d49dc8-kqrxd\" (UID: \"692fd2c9-5168-48fb-9909-7b1cab678b75\") " pod="openstack/keystone-6c78d49dc8-kqrxd" Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.228959 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/692fd2c9-5168-48fb-9909-7b1cab678b75-combined-ca-bundle\") pod \"keystone-6c78d49dc8-kqrxd\" (UID: \"692fd2c9-5168-48fb-9909-7b1cab678b75\") " pod="openstack/keystone-6c78d49dc8-kqrxd" Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.229016 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/692fd2c9-5168-48fb-9909-7b1cab678b75-internal-tls-certs\") pod \"keystone-6c78d49dc8-kqrxd\" (UID: \"692fd2c9-5168-48fb-9909-7b1cab678b75\") " pod="openstack/keystone-6c78d49dc8-kqrxd" Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.239079 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/692fd2c9-5168-48fb-9909-7b1cab678b75-combined-ca-bundle\") pod \"keystone-6c78d49dc8-kqrxd\" (UID: \"692fd2c9-5168-48fb-9909-7b1cab678b75\") " pod="openstack/keystone-6c78d49dc8-kqrxd" Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.239447 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/692fd2c9-5168-48fb-9909-7b1cab678b75-config-data\") pod \"keystone-6c78d49dc8-kqrxd\" (UID: \"692fd2c9-5168-48fb-9909-7b1cab678b75\") " pod="openstack/keystone-6c78d49dc8-kqrxd" Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.244687 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/692fd2c9-5168-48fb-9909-7b1cab678b75-fernet-keys\") pod \"keystone-6c78d49dc8-kqrxd\" (UID: \"692fd2c9-5168-48fb-9909-7b1cab678b75\") " pod="openstack/keystone-6c78d49dc8-kqrxd" Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.245071 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/692fd2c9-5168-48fb-9909-7b1cab678b75-credential-keys\") pod \"keystone-6c78d49dc8-kqrxd\" (UID: \"692fd2c9-5168-48fb-9909-7b1cab678b75\") " pod="openstack/keystone-6c78d49dc8-kqrxd" Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.245357 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/692fd2c9-5168-48fb-9909-7b1cab678b75-scripts\") pod \"keystone-6c78d49dc8-kqrxd\" (UID: \"692fd2c9-5168-48fb-9909-7b1cab678b75\") " pod="openstack/keystone-6c78d49dc8-kqrxd" Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.246156 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/692fd2c9-5168-48fb-9909-7b1cab678b75-internal-tls-certs\") pod \"keystone-6c78d49dc8-kqrxd\" (UID: \"692fd2c9-5168-48fb-9909-7b1cab678b75\") " pod="openstack/keystone-6c78d49dc8-kqrxd" Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.246478 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/692fd2c9-5168-48fb-9909-7b1cab678b75-public-tls-certs\") pod \"keystone-6c78d49dc8-kqrxd\" (UID: \"692fd2c9-5168-48fb-9909-7b1cab678b75\") " pod="openstack/keystone-6c78d49dc8-kqrxd" Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.256211 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8l6km\" (UniqueName: \"kubernetes.io/projected/692fd2c9-5168-48fb-9909-7b1cab678b75-kube-api-access-8l6km\") pod \"keystone-6c78d49dc8-kqrxd\" (UID: \"692fd2c9-5168-48fb-9909-7b1cab678b75\") " pod="openstack/keystone-6c78d49dc8-kqrxd" Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.401916 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6c78d49dc8-kqrxd" Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.537484 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e4a730c4-623d-4ced-a687-bec2d627eae9","Type":"ContainerStarted","Data":"f2ce0409cad8baf6e2ed54568f34ac02c6b59d18905648cee072364d2def57b6"} Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.539470 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-zvfqz" event={"ID":"5c9be753-da64-4599-b5d1-baf7f9be58d4","Type":"ContainerStarted","Data":"5af93f66b8fa19a435a55e198257289a46272a26e52aa7cc585e6a87d507e6ce"} Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.539700 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-84b966f6c9-zvfqz" Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.544627 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c499ea9a-7047-493b-a529-1b5115119b25","Type":"ContainerStarted","Data":"03648621acfcfc360502f4b3af076777d7d5f8a4f30a001515922855a9ab1fe9"} Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.544659 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c499ea9a-7047-493b-a529-1b5115119b25","Type":"ContainerStarted","Data":"fab1facb1e72c8f3a2d0b0dd31e3b630edb5c37b802f0c642c83a48db6158c1e"} Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.547005 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-658bb978b7-lrh6p" event={"ID":"e8e55cd0-dac0-402d-beb7-6482894554fb","Type":"ContainerStarted","Data":"b5130db8b3dc0e1225a5d2acf85eae1e358aecaca0980c214c42495a2a2587f6"} Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.547025 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-658bb978b7-lrh6p" event={"ID":"e8e55cd0-dac0-402d-beb7-6482894554fb","Type":"ContainerStarted","Data":"f297b1aee7fa34dbaa63aca4cb0a1184777fa295dc43f2c7fd0dd78feea58317"} Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.547791 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-658bb978b7-lrh6p" Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.578985 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-84b966f6c9-zvfqz" podStartSLOduration=8.578764821 podStartE2EDuration="8.578764821s" podCreationTimestamp="2025-09-30 14:13:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:13:16.559756726 +0000 UTC m=+1025.188843139" watchObservedRunningTime="2025-09-30 14:13:16.578764821 +0000 UTC m=+1025.207851244" Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.583564 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-658bb978b7-lrh6p" podStartSLOduration=6.583533766 podStartE2EDuration="6.583533766s" podCreationTimestamp="2025-09-30 14:13:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:13:16.581310248 +0000 UTC m=+1025.210396671" watchObservedRunningTime="2025-09-30 14:13:16.583533766 +0000 UTC m=+1025.212620189" Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.728420 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 14:13:16 crc kubenswrapper[4840]: I0930 14:13:16.901177 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6c78d49dc8-kqrxd"] Sep 30 14:13:17 crc kubenswrapper[4840]: I0930 14:13:17.564476 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c499ea9a-7047-493b-a529-1b5115119b25","Type":"ContainerStarted","Data":"b203574d1cba5c9043bdf32ea9de8523a03e3e11b59553bd30a8431865d24f1d"} Sep 30 14:13:17 crc kubenswrapper[4840]: I0930 14:13:17.566174 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6c78d49dc8-kqrxd" event={"ID":"692fd2c9-5168-48fb-9909-7b1cab678b75","Type":"ContainerStarted","Data":"aaa7acea54d711bcf475c75610be186492875a6246cc495f21a93f72eaef9631"} Sep 30 14:13:17 crc kubenswrapper[4840]: I0930 14:13:17.566199 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6c78d49dc8-kqrxd" event={"ID":"692fd2c9-5168-48fb-9909-7b1cab678b75","Type":"ContainerStarted","Data":"fa3e67f7e803f509a1329b67afe2f4e742ee2c36a77cb2cfa1f32520d40a157b"} Sep 30 14:13:17 crc kubenswrapper[4840]: I0930 14:13:17.566580 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-6c78d49dc8-kqrxd" Sep 30 14:13:17 crc kubenswrapper[4840]: I0930 14:13:17.567767 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b96c744c-10e9-4c0c-9ece-069c0c90e67a","Type":"ContainerStarted","Data":"f19c53a63f714f8e0a84ed1833fdf7a816a89d7939ff5aad1ade067169e78a35"} Sep 30 14:13:17 crc kubenswrapper[4840]: I0930 14:13:17.567967 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b96c744c-10e9-4c0c-9ece-069c0c90e67a","Type":"ContainerStarted","Data":"3418dbe2ddffdf60c07068d9f73484617a103469d7d8c7af4840e35d4d00c760"} Sep 30 14:13:17 crc kubenswrapper[4840]: I0930 14:13:17.597572 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.597538249 podStartE2EDuration="7.597538249s" podCreationTimestamp="2025-09-30 14:13:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:13:17.597068257 +0000 UTC m=+1026.226154680" watchObservedRunningTime="2025-09-30 14:13:17.597538249 +0000 UTC m=+1026.226624682" Sep 30 14:13:18 crc kubenswrapper[4840]: I0930 14:13:18.576312 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9hcwn" event={"ID":"2c465939-fcbe-44b2-b7ae-a24b5a581dbc","Type":"ContainerStarted","Data":"7b37e879a3a49adb8414c595ca1aea847737f1f8f450957b4c08d0a2f64813e5"} Sep 30 14:13:18 crc kubenswrapper[4840]: I0930 14:13:18.578035 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b96c744c-10e9-4c0c-9ece-069c0c90e67a","Type":"ContainerStarted","Data":"3c95288a14cc91765a03616025aa0e85a31db07cc1cc694550d048ce584c730b"} Sep 30 14:13:18 crc kubenswrapper[4840]: I0930 14:13:18.594330 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-6c78d49dc8-kqrxd" podStartSLOduration=3.594313283 podStartE2EDuration="3.594313283s" podCreationTimestamp="2025-09-30 14:13:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:13:17.630255322 +0000 UTC m=+1026.259341745" watchObservedRunningTime="2025-09-30 14:13:18.594313283 +0000 UTC m=+1027.223399696" Sep 30 14:13:18 crc kubenswrapper[4840]: I0930 14:13:18.596861 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5595d79c68-dgsth" podUID="5e9b87b2-4836-4857-8514-b2dcfe4de1f5" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Sep 30 14:13:18 crc kubenswrapper[4840]: I0930 14:13:18.620156 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-9hcwn" podStartSLOduration=1.694290658 podStartE2EDuration="35.620140477s" podCreationTimestamp="2025-09-30 14:12:43 +0000 UTC" firstStartedPulling="2025-09-30 14:12:43.695641735 +0000 UTC m=+992.324728158" lastFinishedPulling="2025-09-30 14:13:17.621491554 +0000 UTC m=+1026.250577977" observedRunningTime="2025-09-30 14:13:18.589999821 +0000 UTC m=+1027.219086244" watchObservedRunningTime="2025-09-30 14:13:18.620140477 +0000 UTC m=+1027.249226900" Sep 30 14:13:18 crc kubenswrapper[4840]: I0930 14:13:18.621139 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.621132573 podStartE2EDuration="3.621132573s" podCreationTimestamp="2025-09-30 14:13:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:13:18.619940912 +0000 UTC m=+1027.249027335" watchObservedRunningTime="2025-09-30 14:13:18.621132573 +0000 UTC m=+1027.250218996" Sep 30 14:13:18 crc kubenswrapper[4840]: I0930 14:13:18.733051 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-94dfcc48d-mlrjd" podUID="1138b51b-edf9-42d2-8ea4-4d4b9122eae4" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.150:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.150:8443: connect: connection refused" Sep 30 14:13:20 crc kubenswrapper[4840]: I0930 14:13:20.599103 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-sp7q5" event={"ID":"ad748b83-4edd-4545-9386-032185c531c5","Type":"ContainerStarted","Data":"8391b03af598c7e9580a0e3ab6d61f6d82ad2ca29aee7293a25f1fb098fad220"} Sep 30 14:13:20 crc kubenswrapper[4840]: I0930 14:13:20.620698 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-sp7q5" podStartSLOduration=2.958615335 podStartE2EDuration="41.620681167s" podCreationTimestamp="2025-09-30 14:12:39 +0000 UTC" firstStartedPulling="2025-09-30 14:12:40.908209213 +0000 UTC m=+989.537295636" lastFinishedPulling="2025-09-30 14:13:19.570275045 +0000 UTC m=+1028.199361468" observedRunningTime="2025-09-30 14:13:20.617638397 +0000 UTC m=+1029.246724820" watchObservedRunningTime="2025-09-30 14:13:20.620681167 +0000 UTC m=+1029.249767580" Sep 30 14:13:20 crc kubenswrapper[4840]: I0930 14:13:20.783734 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Sep 30 14:13:20 crc kubenswrapper[4840]: I0930 14:13:20.783903 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Sep 30 14:13:20 crc kubenswrapper[4840]: I0930 14:13:20.840178 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Sep 30 14:13:20 crc kubenswrapper[4840]: I0930 14:13:20.849333 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Sep 30 14:13:21 crc kubenswrapper[4840]: I0930 14:13:21.610467 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-9vd76" event={"ID":"242d13bb-6113-421b-83d1-fcccf91a9c68","Type":"ContainerStarted","Data":"58f6da62f61ae533f5086dfb033f031419373cb17ddf4e1ee41d20d765c50f2c"} Sep 30 14:13:21 crc kubenswrapper[4840]: I0930 14:13:21.612754 4840 generic.go:334] "Generic (PLEG): container finished" podID="2c465939-fcbe-44b2-b7ae-a24b5a581dbc" containerID="7b37e879a3a49adb8414c595ca1aea847737f1f8f450957b4c08d0a2f64813e5" exitCode=0 Sep 30 14:13:21 crc kubenswrapper[4840]: I0930 14:13:21.613825 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9hcwn" event={"ID":"2c465939-fcbe-44b2-b7ae-a24b5a581dbc","Type":"ContainerDied","Data":"7b37e879a3a49adb8414c595ca1aea847737f1f8f450957b4c08d0a2f64813e5"} Sep 30 14:13:21 crc kubenswrapper[4840]: I0930 14:13:21.613875 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Sep 30 14:13:21 crc kubenswrapper[4840]: I0930 14:13:21.614140 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Sep 30 14:13:21 crc kubenswrapper[4840]: I0930 14:13:21.637850 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-9vd76" podStartSLOduration=2.412018222 podStartE2EDuration="42.637836712s" podCreationTimestamp="2025-09-30 14:12:39 +0000 UTC" firstStartedPulling="2025-09-30 14:12:40.602323247 +0000 UTC m=+989.231409670" lastFinishedPulling="2025-09-30 14:13:20.828141737 +0000 UTC m=+1029.457228160" observedRunningTime="2025-09-30 14:13:21.636489397 +0000 UTC m=+1030.265575820" watchObservedRunningTime="2025-09-30 14:13:21.637836712 +0000 UTC m=+1030.266923135" Sep 30 14:13:21 crc kubenswrapper[4840]: I0930 14:13:21.871780 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:13:21 crc kubenswrapper[4840]: I0930 14:13:21.872535 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:13:22 crc kubenswrapper[4840]: I0930 14:13:22.622757 4840 generic.go:334] "Generic (PLEG): container finished" podID="ad748b83-4edd-4545-9386-032185c531c5" containerID="8391b03af598c7e9580a0e3ab6d61f6d82ad2ca29aee7293a25f1fb098fad220" exitCode=0 Sep 30 14:13:22 crc kubenswrapper[4840]: I0930 14:13:22.622836 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-sp7q5" event={"ID":"ad748b83-4edd-4545-9386-032185c531c5","Type":"ContainerDied","Data":"8391b03af598c7e9580a0e3ab6d61f6d82ad2ca29aee7293a25f1fb098fad220"} Sep 30 14:13:23 crc kubenswrapper[4840]: I0930 14:13:23.631091 4840 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 14:13:23 crc kubenswrapper[4840]: I0930 14:13:23.804809 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-84b966f6c9-zvfqz" Sep 30 14:13:23 crc kubenswrapper[4840]: I0930 14:13:23.862399 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-gxf58"] Sep 30 14:13:23 crc kubenswrapper[4840]: I0930 14:13:23.862672 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8b5c85b87-gxf58" podUID="b73e88ee-cb25-466d-9860-0dd54a1c1f4c" containerName="dnsmasq-dns" containerID="cri-o://8f1dc45f77092a5778340c020c0952b71db3f7fad8ff85b132ad5c8cc6e7a5c5" gracePeriod=10 Sep 30 14:13:24 crc kubenswrapper[4840]: I0930 14:13:24.029049 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Sep 30 14:13:24 crc kubenswrapper[4840]: I0930 14:13:24.030815 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Sep 30 14:13:24 crc kubenswrapper[4840]: I0930 14:13:24.654907 4840 generic.go:334] "Generic (PLEG): container finished" podID="b73e88ee-cb25-466d-9860-0dd54a1c1f4c" containerID="8f1dc45f77092a5778340c020c0952b71db3f7fad8ff85b132ad5c8cc6e7a5c5" exitCode=0 Sep 30 14:13:24 crc kubenswrapper[4840]: I0930 14:13:24.655004 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-gxf58" event={"ID":"b73e88ee-cb25-466d-9860-0dd54a1c1f4c","Type":"ContainerDied","Data":"8f1dc45f77092a5778340c020c0952b71db3f7fad8ff85b132ad5c8cc6e7a5c5"} Sep 30 14:13:24 crc kubenswrapper[4840]: I0930 14:13:24.781912 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-8b5c85b87-gxf58" podUID="b73e88ee-cb25-466d-9860-0dd54a1c1f4c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.140:5353: connect: connection refused" Sep 30 14:13:25 crc kubenswrapper[4840]: I0930 14:13:25.935508 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Sep 30 14:13:25 crc kubenswrapper[4840]: I0930 14:13:25.935775 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Sep 30 14:13:25 crc kubenswrapper[4840]: I0930 14:13:25.979262 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Sep 30 14:13:26 crc kubenswrapper[4840]: I0930 14:13:26.000263 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Sep 30 14:13:26 crc kubenswrapper[4840]: I0930 14:13:26.594791 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9hcwn" Sep 30 14:13:26 crc kubenswrapper[4840]: I0930 14:13:26.599003 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-sp7q5" Sep 30 14:13:26 crc kubenswrapper[4840]: I0930 14:13:26.682406 4840 generic.go:334] "Generic (PLEG): container finished" podID="242d13bb-6113-421b-83d1-fcccf91a9c68" containerID="58f6da62f61ae533f5086dfb033f031419373cb17ddf4e1ee41d20d765c50f2c" exitCode=0 Sep 30 14:13:26 crc kubenswrapper[4840]: I0930 14:13:26.682442 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-9vd76" event={"ID":"242d13bb-6113-421b-83d1-fcccf91a9c68","Type":"ContainerDied","Data":"58f6da62f61ae533f5086dfb033f031419373cb17ddf4e1ee41d20d765c50f2c"} Sep 30 14:13:26 crc kubenswrapper[4840]: I0930 14:13:26.685939 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9hcwn" event={"ID":"2c465939-fcbe-44b2-b7ae-a24b5a581dbc","Type":"ContainerDied","Data":"3958b3f15ef2941ffd7ae2a69bdb6a938691349e2395edb10287002e93dd7857"} Sep 30 14:13:26 crc kubenswrapper[4840]: I0930 14:13:26.685973 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3958b3f15ef2941ffd7ae2a69bdb6a938691349e2395edb10287002e93dd7857" Sep 30 14:13:26 crc kubenswrapper[4840]: I0930 14:13:26.686021 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9hcwn" Sep 30 14:13:26 crc kubenswrapper[4840]: I0930 14:13:26.693414 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-sp7q5" event={"ID":"ad748b83-4edd-4545-9386-032185c531c5","Type":"ContainerDied","Data":"82168598b29a794bbf74c3fee4e2676ee2b702525cf44ada0f7498f39206978d"} Sep 30 14:13:26 crc kubenswrapper[4840]: I0930 14:13:26.693864 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="82168598b29a794bbf74c3fee4e2676ee2b702525cf44ada0f7498f39206978d" Sep 30 14:13:26 crc kubenswrapper[4840]: I0930 14:13:26.693891 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Sep 30 14:13:26 crc kubenswrapper[4840]: I0930 14:13:26.693923 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Sep 30 14:13:26 crc kubenswrapper[4840]: I0930 14:13:26.694238 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-sp7q5" Sep 30 14:13:26 crc kubenswrapper[4840]: I0930 14:13:26.756580 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4l6jg\" (UniqueName: \"kubernetes.io/projected/2c465939-fcbe-44b2-b7ae-a24b5a581dbc-kube-api-access-4l6jg\") pod \"2c465939-fcbe-44b2-b7ae-a24b5a581dbc\" (UID: \"2c465939-fcbe-44b2-b7ae-a24b5a581dbc\") " Sep 30 14:13:26 crc kubenswrapper[4840]: I0930 14:13:26.756672 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2c465939-fcbe-44b2-b7ae-a24b5a581dbc-db-sync-config-data\") pod \"2c465939-fcbe-44b2-b7ae-a24b5a581dbc\" (UID: \"2c465939-fcbe-44b2-b7ae-a24b5a581dbc\") " Sep 30 14:13:26 crc kubenswrapper[4840]: I0930 14:13:26.756710 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c465939-fcbe-44b2-b7ae-a24b5a581dbc-combined-ca-bundle\") pod \"2c465939-fcbe-44b2-b7ae-a24b5a581dbc\" (UID: \"2c465939-fcbe-44b2-b7ae-a24b5a581dbc\") " Sep 30 14:13:26 crc kubenswrapper[4840]: I0930 14:13:26.756754 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ad748b83-4edd-4545-9386-032185c531c5-scripts\") pod \"ad748b83-4edd-4545-9386-032185c531c5\" (UID: \"ad748b83-4edd-4545-9386-032185c531c5\") " Sep 30 14:13:26 crc kubenswrapper[4840]: I0930 14:13:26.756778 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad748b83-4edd-4545-9386-032185c531c5-config-data\") pod \"ad748b83-4edd-4545-9386-032185c531c5\" (UID: \"ad748b83-4edd-4545-9386-032185c531c5\") " Sep 30 14:13:26 crc kubenswrapper[4840]: I0930 14:13:26.756873 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7g529\" (UniqueName: \"kubernetes.io/projected/ad748b83-4edd-4545-9386-032185c531c5-kube-api-access-7g529\") pod \"ad748b83-4edd-4545-9386-032185c531c5\" (UID: \"ad748b83-4edd-4545-9386-032185c531c5\") " Sep 30 14:13:26 crc kubenswrapper[4840]: I0930 14:13:26.756898 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad748b83-4edd-4545-9386-032185c531c5-logs\") pod \"ad748b83-4edd-4545-9386-032185c531c5\" (UID: \"ad748b83-4edd-4545-9386-032185c531c5\") " Sep 30 14:13:26 crc kubenswrapper[4840]: I0930 14:13:26.756924 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad748b83-4edd-4545-9386-032185c531c5-combined-ca-bundle\") pod \"ad748b83-4edd-4545-9386-032185c531c5\" (UID: \"ad748b83-4edd-4545-9386-032185c531c5\") " Sep 30 14:13:26 crc kubenswrapper[4840]: I0930 14:13:26.763887 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad748b83-4edd-4545-9386-032185c531c5-kube-api-access-7g529" (OuterVolumeSpecName: "kube-api-access-7g529") pod "ad748b83-4edd-4545-9386-032185c531c5" (UID: "ad748b83-4edd-4545-9386-032185c531c5"). InnerVolumeSpecName "kube-api-access-7g529". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:13:26 crc kubenswrapper[4840]: I0930 14:13:26.765283 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad748b83-4edd-4545-9386-032185c531c5-scripts" (OuterVolumeSpecName: "scripts") pod "ad748b83-4edd-4545-9386-032185c531c5" (UID: "ad748b83-4edd-4545-9386-032185c531c5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:26 crc kubenswrapper[4840]: I0930 14:13:26.765804 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad748b83-4edd-4545-9386-032185c531c5-logs" (OuterVolumeSpecName: "logs") pod "ad748b83-4edd-4545-9386-032185c531c5" (UID: "ad748b83-4edd-4545-9386-032185c531c5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:13:26 crc kubenswrapper[4840]: I0930 14:13:26.766852 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c465939-fcbe-44b2-b7ae-a24b5a581dbc-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "2c465939-fcbe-44b2-b7ae-a24b5a581dbc" (UID: "2c465939-fcbe-44b2-b7ae-a24b5a581dbc"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:26 crc kubenswrapper[4840]: I0930 14:13:26.784034 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c465939-fcbe-44b2-b7ae-a24b5a581dbc-kube-api-access-4l6jg" (OuterVolumeSpecName: "kube-api-access-4l6jg") pod "2c465939-fcbe-44b2-b7ae-a24b5a581dbc" (UID: "2c465939-fcbe-44b2-b7ae-a24b5a581dbc"). InnerVolumeSpecName "kube-api-access-4l6jg". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:13:26 crc kubenswrapper[4840]: I0930 14:13:26.795790 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad748b83-4edd-4545-9386-032185c531c5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ad748b83-4edd-4545-9386-032185c531c5" (UID: "ad748b83-4edd-4545-9386-032185c531c5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:26 crc kubenswrapper[4840]: I0930 14:13:26.816868 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c465939-fcbe-44b2-b7ae-a24b5a581dbc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2c465939-fcbe-44b2-b7ae-a24b5a581dbc" (UID: "2c465939-fcbe-44b2-b7ae-a24b5a581dbc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:26 crc kubenswrapper[4840]: I0930 14:13:26.842003 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad748b83-4edd-4545-9386-032185c531c5-config-data" (OuterVolumeSpecName: "config-data") pod "ad748b83-4edd-4545-9386-032185c531c5" (UID: "ad748b83-4edd-4545-9386-032185c531c5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:26 crc kubenswrapper[4840]: I0930 14:13:26.859380 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c465939-fcbe-44b2-b7ae-a24b5a581dbc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:26 crc kubenswrapper[4840]: I0930 14:13:26.859417 4840 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ad748b83-4edd-4545-9386-032185c531c5-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:26 crc kubenswrapper[4840]: I0930 14:13:26.859430 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad748b83-4edd-4545-9386-032185c531c5-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:26 crc kubenswrapper[4840]: I0930 14:13:26.859442 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7g529\" (UniqueName: \"kubernetes.io/projected/ad748b83-4edd-4545-9386-032185c531c5-kube-api-access-7g529\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:26 crc kubenswrapper[4840]: I0930 14:13:26.859454 4840 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad748b83-4edd-4545-9386-032185c531c5-logs\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:26 crc kubenswrapper[4840]: I0930 14:13:26.859464 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad748b83-4edd-4545-9386-032185c531c5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:26 crc kubenswrapper[4840]: I0930 14:13:26.859474 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4l6jg\" (UniqueName: \"kubernetes.io/projected/2c465939-fcbe-44b2-b7ae-a24b5a581dbc-kube-api-access-4l6jg\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:26 crc kubenswrapper[4840]: I0930 14:13:26.859486 4840 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2c465939-fcbe-44b2-b7ae-a24b5a581dbc-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:27 crc kubenswrapper[4840]: I0930 14:13:27.807139 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-6cbd5fdcfb-hn4pn"] Sep 30 14:13:27 crc kubenswrapper[4840]: E0930 14:13:27.807872 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad748b83-4edd-4545-9386-032185c531c5" containerName="placement-db-sync" Sep 30 14:13:27 crc kubenswrapper[4840]: I0930 14:13:27.807884 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad748b83-4edd-4545-9386-032185c531c5" containerName="placement-db-sync" Sep 30 14:13:27 crc kubenswrapper[4840]: E0930 14:13:27.807916 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c465939-fcbe-44b2-b7ae-a24b5a581dbc" containerName="barbican-db-sync" Sep 30 14:13:27 crc kubenswrapper[4840]: I0930 14:13:27.807921 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c465939-fcbe-44b2-b7ae-a24b5a581dbc" containerName="barbican-db-sync" Sep 30 14:13:27 crc kubenswrapper[4840]: I0930 14:13:27.808078 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c465939-fcbe-44b2-b7ae-a24b5a581dbc" containerName="barbican-db-sync" Sep 30 14:13:27 crc kubenswrapper[4840]: I0930 14:13:27.808093 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad748b83-4edd-4545-9386-032185c531c5" containerName="placement-db-sync" Sep 30 14:13:27 crc kubenswrapper[4840]: I0930 14:13:27.809005 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6cbd5fdcfb-hn4pn" Sep 30 14:13:27 crc kubenswrapper[4840]: I0930 14:13:27.817228 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Sep 30 14:13:27 crc kubenswrapper[4840]: I0930 14:13:27.817481 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Sep 30 14:13:27 crc kubenswrapper[4840]: I0930 14:13:27.817345 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Sep 30 14:13:27 crc kubenswrapper[4840]: I0930 14:13:27.817767 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Sep 30 14:13:27 crc kubenswrapper[4840]: I0930 14:13:27.817400 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-xq6vq" Sep 30 14:13:27 crc kubenswrapper[4840]: I0930 14:13:27.936878 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6cbd5fdcfb-hn4pn"] Sep 30 14:13:27 crc kubenswrapper[4840]: I0930 14:13:27.984090 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-668f8fbd45-fpqbx"] Sep 30 14:13:27 crc kubenswrapper[4840]: I0930 14:13:27.985534 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-668f8fbd45-fpqbx" Sep 30 14:13:27 crc kubenswrapper[4840]: I0930 14:13:27.988475 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1eaca45-6337-4f35-ac67-dfa0fc83e1bb-public-tls-certs\") pod \"placement-6cbd5fdcfb-hn4pn\" (UID: \"c1eaca45-6337-4f35-ac67-dfa0fc83e1bb\") " pod="openstack/placement-6cbd5fdcfb-hn4pn" Sep 30 14:13:27 crc kubenswrapper[4840]: I0930 14:13:27.988542 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwhd8\" (UniqueName: \"kubernetes.io/projected/c1eaca45-6337-4f35-ac67-dfa0fc83e1bb-kube-api-access-cwhd8\") pod \"placement-6cbd5fdcfb-hn4pn\" (UID: \"c1eaca45-6337-4f35-ac67-dfa0fc83e1bb\") " pod="openstack/placement-6cbd5fdcfb-hn4pn" Sep 30 14:13:27 crc kubenswrapper[4840]: I0930 14:13:27.988594 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1eaca45-6337-4f35-ac67-dfa0fc83e1bb-config-data\") pod \"placement-6cbd5fdcfb-hn4pn\" (UID: \"c1eaca45-6337-4f35-ac67-dfa0fc83e1bb\") " pod="openstack/placement-6cbd5fdcfb-hn4pn" Sep 30 14:13:27 crc kubenswrapper[4840]: I0930 14:13:27.988618 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1eaca45-6337-4f35-ac67-dfa0fc83e1bb-scripts\") pod \"placement-6cbd5fdcfb-hn4pn\" (UID: \"c1eaca45-6337-4f35-ac67-dfa0fc83e1bb\") " pod="openstack/placement-6cbd5fdcfb-hn4pn" Sep 30 14:13:27 crc kubenswrapper[4840]: I0930 14:13:27.988667 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1eaca45-6337-4f35-ac67-dfa0fc83e1bb-combined-ca-bundle\") pod \"placement-6cbd5fdcfb-hn4pn\" (UID: \"c1eaca45-6337-4f35-ac67-dfa0fc83e1bb\") " pod="openstack/placement-6cbd5fdcfb-hn4pn" Sep 30 14:13:27 crc kubenswrapper[4840]: I0930 14:13:27.988694 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1eaca45-6337-4f35-ac67-dfa0fc83e1bb-internal-tls-certs\") pod \"placement-6cbd5fdcfb-hn4pn\" (UID: \"c1eaca45-6337-4f35-ac67-dfa0fc83e1bb\") " pod="openstack/placement-6cbd5fdcfb-hn4pn" Sep 30 14:13:27 crc kubenswrapper[4840]: I0930 14:13:27.988731 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1eaca45-6337-4f35-ac67-dfa0fc83e1bb-logs\") pod \"placement-6cbd5fdcfb-hn4pn\" (UID: \"c1eaca45-6337-4f35-ac67-dfa0fc83e1bb\") " pod="openstack/placement-6cbd5fdcfb-hn4pn" Sep 30 14:13:27 crc kubenswrapper[4840]: I0930 14:13:27.989208 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Sep 30 14:13:27 crc kubenswrapper[4840]: I0930 14:13:27.989366 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Sep 30 14:13:27 crc kubenswrapper[4840]: I0930 14:13:27.989464 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-g5svs" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.027627 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-668f8fbd45-fpqbx"] Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.077665 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-67bf9f4b94-nk9h8"] Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.079243 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-67bf9f4b94-nk9h8" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.086816 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.097450 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1eaca45-6337-4f35-ac67-dfa0fc83e1bb-logs\") pod \"placement-6cbd5fdcfb-hn4pn\" (UID: \"c1eaca45-6337-4f35-ac67-dfa0fc83e1bb\") " pod="openstack/placement-6cbd5fdcfb-hn4pn" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.097496 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fa943f1-8698-41e3-a893-fd1c8845076a-config-data\") pod \"barbican-worker-668f8fbd45-fpqbx\" (UID: \"0fa943f1-8698-41e3-a893-fd1c8845076a\") " pod="openstack/barbican-worker-668f8fbd45-fpqbx" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.097533 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1eaca45-6337-4f35-ac67-dfa0fc83e1bb-public-tls-certs\") pod \"placement-6cbd5fdcfb-hn4pn\" (UID: \"c1eaca45-6337-4f35-ac67-dfa0fc83e1bb\") " pod="openstack/placement-6cbd5fdcfb-hn4pn" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.097592 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwhd8\" (UniqueName: \"kubernetes.io/projected/c1eaca45-6337-4f35-ac67-dfa0fc83e1bb-kube-api-access-cwhd8\") pod \"placement-6cbd5fdcfb-hn4pn\" (UID: \"c1eaca45-6337-4f35-ac67-dfa0fc83e1bb\") " pod="openstack/placement-6cbd5fdcfb-hn4pn" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.097625 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1eaca45-6337-4f35-ac67-dfa0fc83e1bb-config-data\") pod \"placement-6cbd5fdcfb-hn4pn\" (UID: \"c1eaca45-6337-4f35-ac67-dfa0fc83e1bb\") " pod="openstack/placement-6cbd5fdcfb-hn4pn" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.097643 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0fa943f1-8698-41e3-a893-fd1c8845076a-config-data-custom\") pod \"barbican-worker-668f8fbd45-fpqbx\" (UID: \"0fa943f1-8698-41e3-a893-fd1c8845076a\") " pod="openstack/barbican-worker-668f8fbd45-fpqbx" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.097667 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1eaca45-6337-4f35-ac67-dfa0fc83e1bb-scripts\") pod \"placement-6cbd5fdcfb-hn4pn\" (UID: \"c1eaca45-6337-4f35-ac67-dfa0fc83e1bb\") " pod="openstack/placement-6cbd5fdcfb-hn4pn" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.097691 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fa943f1-8698-41e3-a893-fd1c8845076a-combined-ca-bundle\") pod \"barbican-worker-668f8fbd45-fpqbx\" (UID: \"0fa943f1-8698-41e3-a893-fd1c8845076a\") " pod="openstack/barbican-worker-668f8fbd45-fpqbx" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.097730 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nndfc\" (UniqueName: \"kubernetes.io/projected/0fa943f1-8698-41e3-a893-fd1c8845076a-kube-api-access-nndfc\") pod \"barbican-worker-668f8fbd45-fpqbx\" (UID: \"0fa943f1-8698-41e3-a893-fd1c8845076a\") " pod="openstack/barbican-worker-668f8fbd45-fpqbx" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.097757 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1eaca45-6337-4f35-ac67-dfa0fc83e1bb-combined-ca-bundle\") pod \"placement-6cbd5fdcfb-hn4pn\" (UID: \"c1eaca45-6337-4f35-ac67-dfa0fc83e1bb\") " pod="openstack/placement-6cbd5fdcfb-hn4pn" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.097781 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1eaca45-6337-4f35-ac67-dfa0fc83e1bb-internal-tls-certs\") pod \"placement-6cbd5fdcfb-hn4pn\" (UID: \"c1eaca45-6337-4f35-ac67-dfa0fc83e1bb\") " pod="openstack/placement-6cbd5fdcfb-hn4pn" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.097800 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fa943f1-8698-41e3-a893-fd1c8845076a-logs\") pod \"barbican-worker-668f8fbd45-fpqbx\" (UID: \"0fa943f1-8698-41e3-a893-fd1c8845076a\") " pod="openstack/barbican-worker-668f8fbd45-fpqbx" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.098180 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1eaca45-6337-4f35-ac67-dfa0fc83e1bb-logs\") pod \"placement-6cbd5fdcfb-hn4pn\" (UID: \"c1eaca45-6337-4f35-ac67-dfa0fc83e1bb\") " pod="openstack/placement-6cbd5fdcfb-hn4pn" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.103674 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-67bf9f4b94-nk9h8"] Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.124269 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1eaca45-6337-4f35-ac67-dfa0fc83e1bb-public-tls-certs\") pod \"placement-6cbd5fdcfb-hn4pn\" (UID: \"c1eaca45-6337-4f35-ac67-dfa0fc83e1bb\") " pod="openstack/placement-6cbd5fdcfb-hn4pn" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.131149 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1eaca45-6337-4f35-ac67-dfa0fc83e1bb-scripts\") pod \"placement-6cbd5fdcfb-hn4pn\" (UID: \"c1eaca45-6337-4f35-ac67-dfa0fc83e1bb\") " pod="openstack/placement-6cbd5fdcfb-hn4pn" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.131187 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1eaca45-6337-4f35-ac67-dfa0fc83e1bb-config-data\") pod \"placement-6cbd5fdcfb-hn4pn\" (UID: \"c1eaca45-6337-4f35-ac67-dfa0fc83e1bb\") " pod="openstack/placement-6cbd5fdcfb-hn4pn" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.140102 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1eaca45-6337-4f35-ac67-dfa0fc83e1bb-internal-tls-certs\") pod \"placement-6cbd5fdcfb-hn4pn\" (UID: \"c1eaca45-6337-4f35-ac67-dfa0fc83e1bb\") " pod="openstack/placement-6cbd5fdcfb-hn4pn" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.151765 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1eaca45-6337-4f35-ac67-dfa0fc83e1bb-combined-ca-bundle\") pod \"placement-6cbd5fdcfb-hn4pn\" (UID: \"c1eaca45-6337-4f35-ac67-dfa0fc83e1bb\") " pod="openstack/placement-6cbd5fdcfb-hn4pn" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.158316 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwhd8\" (UniqueName: \"kubernetes.io/projected/c1eaca45-6337-4f35-ac67-dfa0fc83e1bb-kube-api-access-cwhd8\") pod \"placement-6cbd5fdcfb-hn4pn\" (UID: \"c1eaca45-6337-4f35-ac67-dfa0fc83e1bb\") " pod="openstack/placement-6cbd5fdcfb-hn4pn" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.172235 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6cbd5fdcfb-hn4pn" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.177473 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-gzfst"] Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.191972 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-gzfst"] Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.192269 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-gzfst" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.204427 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fa943f1-8698-41e3-a893-fd1c8845076a-logs\") pod \"barbican-worker-668f8fbd45-fpqbx\" (UID: \"0fa943f1-8698-41e3-a893-fd1c8845076a\") " pod="openstack/barbican-worker-668f8fbd45-fpqbx" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.204481 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae9a65ac-f1a7-4ef0-aeba-aca57985ae87-config-data\") pod \"barbican-keystone-listener-67bf9f4b94-nk9h8\" (UID: \"ae9a65ac-f1a7-4ef0-aeba-aca57985ae87\") " pod="openstack/barbican-keystone-listener-67bf9f4b94-nk9h8" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.204517 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fa943f1-8698-41e3-a893-fd1c8845076a-config-data\") pod \"barbican-worker-668f8fbd45-fpqbx\" (UID: \"0fa943f1-8698-41e3-a893-fd1c8845076a\") " pod="openstack/barbican-worker-668f8fbd45-fpqbx" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.204541 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae9a65ac-f1a7-4ef0-aeba-aca57985ae87-combined-ca-bundle\") pod \"barbican-keystone-listener-67bf9f4b94-nk9h8\" (UID: \"ae9a65ac-f1a7-4ef0-aeba-aca57985ae87\") " pod="openstack/barbican-keystone-listener-67bf9f4b94-nk9h8" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.218770 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae9a65ac-f1a7-4ef0-aeba-aca57985ae87-logs\") pod \"barbican-keystone-listener-67bf9f4b94-nk9h8\" (UID: \"ae9a65ac-f1a7-4ef0-aeba-aca57985ae87\") " pod="openstack/barbican-keystone-listener-67bf9f4b94-nk9h8" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.218897 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0fa943f1-8698-41e3-a893-fd1c8845076a-config-data-custom\") pod \"barbican-worker-668f8fbd45-fpqbx\" (UID: \"0fa943f1-8698-41e3-a893-fd1c8845076a\") " pod="openstack/barbican-worker-668f8fbd45-fpqbx" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.218926 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fa943f1-8698-41e3-a893-fd1c8845076a-combined-ca-bundle\") pod \"barbican-worker-668f8fbd45-fpqbx\" (UID: \"0fa943f1-8698-41e3-a893-fd1c8845076a\") " pod="openstack/barbican-worker-668f8fbd45-fpqbx" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.218977 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxf9p\" (UniqueName: \"kubernetes.io/projected/ae9a65ac-f1a7-4ef0-aeba-aca57985ae87-kube-api-access-jxf9p\") pod \"barbican-keystone-listener-67bf9f4b94-nk9h8\" (UID: \"ae9a65ac-f1a7-4ef0-aeba-aca57985ae87\") " pod="openstack/barbican-keystone-listener-67bf9f4b94-nk9h8" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.219037 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae9a65ac-f1a7-4ef0-aeba-aca57985ae87-config-data-custom\") pod \"barbican-keystone-listener-67bf9f4b94-nk9h8\" (UID: \"ae9a65ac-f1a7-4ef0-aeba-aca57985ae87\") " pod="openstack/barbican-keystone-listener-67bf9f4b94-nk9h8" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.219067 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nndfc\" (UniqueName: \"kubernetes.io/projected/0fa943f1-8698-41e3-a893-fd1c8845076a-kube-api-access-nndfc\") pod \"barbican-worker-668f8fbd45-fpqbx\" (UID: \"0fa943f1-8698-41e3-a893-fd1c8845076a\") " pod="openstack/barbican-worker-668f8fbd45-fpqbx" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.204983 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fa943f1-8698-41e3-a893-fd1c8845076a-logs\") pod \"barbican-worker-668f8fbd45-fpqbx\" (UID: \"0fa943f1-8698-41e3-a893-fd1c8845076a\") " pod="openstack/barbican-worker-668f8fbd45-fpqbx" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.210540 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fa943f1-8698-41e3-a893-fd1c8845076a-config-data\") pod \"barbican-worker-668f8fbd45-fpqbx\" (UID: \"0fa943f1-8698-41e3-a893-fd1c8845076a\") " pod="openstack/barbican-worker-668f8fbd45-fpqbx" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.230436 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fa943f1-8698-41e3-a893-fd1c8845076a-combined-ca-bundle\") pod \"barbican-worker-668f8fbd45-fpqbx\" (UID: \"0fa943f1-8698-41e3-a893-fd1c8845076a\") " pod="openstack/barbican-worker-668f8fbd45-fpqbx" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.231592 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0fa943f1-8698-41e3-a893-fd1c8845076a-config-data-custom\") pod \"barbican-worker-668f8fbd45-fpqbx\" (UID: \"0fa943f1-8698-41e3-a893-fd1c8845076a\") " pod="openstack/barbican-worker-668f8fbd45-fpqbx" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.245104 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nndfc\" (UniqueName: \"kubernetes.io/projected/0fa943f1-8698-41e3-a893-fd1c8845076a-kube-api-access-nndfc\") pod \"barbican-worker-668f8fbd45-fpqbx\" (UID: \"0fa943f1-8698-41e3-a893-fd1c8845076a\") " pod="openstack/barbican-worker-668f8fbd45-fpqbx" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.259231 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-59645bff48-kxnsb"] Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.260703 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-59645bff48-kxnsb" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.263589 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.289983 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-59645bff48-kxnsb"] Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.322840 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxf9p\" (UniqueName: \"kubernetes.io/projected/ae9a65ac-f1a7-4ef0-aeba-aca57985ae87-kube-api-access-jxf9p\") pod \"barbican-keystone-listener-67bf9f4b94-nk9h8\" (UID: \"ae9a65ac-f1a7-4ef0-aeba-aca57985ae87\") " pod="openstack/barbican-keystone-listener-67bf9f4b94-nk9h8" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.322891 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4t4n\" (UniqueName: \"kubernetes.io/projected/b9947cc2-023b-407f-a6c2-b4b2e2c67455-kube-api-access-x4t4n\") pod \"dnsmasq-dns-75c8ddd69c-gzfst\" (UID: \"b9947cc2-023b-407f-a6c2-b4b2e2c67455\") " pod="openstack/dnsmasq-dns-75c8ddd69c-gzfst" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.322925 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae9a65ac-f1a7-4ef0-aeba-aca57985ae87-config-data-custom\") pod \"barbican-keystone-listener-67bf9f4b94-nk9h8\" (UID: \"ae9a65ac-f1a7-4ef0-aeba-aca57985ae87\") " pod="openstack/barbican-keystone-listener-67bf9f4b94-nk9h8" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.322954 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b9947cc2-023b-407f-a6c2-b4b2e2c67455-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-gzfst\" (UID: \"b9947cc2-023b-407f-a6c2-b4b2e2c67455\") " pod="openstack/dnsmasq-dns-75c8ddd69c-gzfst" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.322987 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9947cc2-023b-407f-a6c2-b4b2e2c67455-config\") pod \"dnsmasq-dns-75c8ddd69c-gzfst\" (UID: \"b9947cc2-023b-407f-a6c2-b4b2e2c67455\") " pod="openstack/dnsmasq-dns-75c8ddd69c-gzfst" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.323038 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae9a65ac-f1a7-4ef0-aeba-aca57985ae87-config-data\") pod \"barbican-keystone-listener-67bf9f4b94-nk9h8\" (UID: \"ae9a65ac-f1a7-4ef0-aeba-aca57985ae87\") " pod="openstack/barbican-keystone-listener-67bf9f4b94-nk9h8" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.323077 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae9a65ac-f1a7-4ef0-aeba-aca57985ae87-combined-ca-bundle\") pod \"barbican-keystone-listener-67bf9f4b94-nk9h8\" (UID: \"ae9a65ac-f1a7-4ef0-aeba-aca57985ae87\") " pod="openstack/barbican-keystone-listener-67bf9f4b94-nk9h8" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.323120 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae9a65ac-f1a7-4ef0-aeba-aca57985ae87-logs\") pod \"barbican-keystone-listener-67bf9f4b94-nk9h8\" (UID: \"ae9a65ac-f1a7-4ef0-aeba-aca57985ae87\") " pod="openstack/barbican-keystone-listener-67bf9f4b94-nk9h8" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.323138 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b9947cc2-023b-407f-a6c2-b4b2e2c67455-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-gzfst\" (UID: \"b9947cc2-023b-407f-a6c2-b4b2e2c67455\") " pod="openstack/dnsmasq-dns-75c8ddd69c-gzfst" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.323153 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b9947cc2-023b-407f-a6c2-b4b2e2c67455-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-gzfst\" (UID: \"b9947cc2-023b-407f-a6c2-b4b2e2c67455\") " pod="openstack/dnsmasq-dns-75c8ddd69c-gzfst" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.323188 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9947cc2-023b-407f-a6c2-b4b2e2c67455-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-gzfst\" (UID: \"b9947cc2-023b-407f-a6c2-b4b2e2c67455\") " pod="openstack/dnsmasq-dns-75c8ddd69c-gzfst" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.324566 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae9a65ac-f1a7-4ef0-aeba-aca57985ae87-logs\") pod \"barbican-keystone-listener-67bf9f4b94-nk9h8\" (UID: \"ae9a65ac-f1a7-4ef0-aeba-aca57985ae87\") " pod="openstack/barbican-keystone-listener-67bf9f4b94-nk9h8" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.329358 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae9a65ac-f1a7-4ef0-aeba-aca57985ae87-config-data-custom\") pod \"barbican-keystone-listener-67bf9f4b94-nk9h8\" (UID: \"ae9a65ac-f1a7-4ef0-aeba-aca57985ae87\") " pod="openstack/barbican-keystone-listener-67bf9f4b94-nk9h8" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.331401 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae9a65ac-f1a7-4ef0-aeba-aca57985ae87-config-data\") pod \"barbican-keystone-listener-67bf9f4b94-nk9h8\" (UID: \"ae9a65ac-f1a7-4ef0-aeba-aca57985ae87\") " pod="openstack/barbican-keystone-listener-67bf9f4b94-nk9h8" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.345329 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae9a65ac-f1a7-4ef0-aeba-aca57985ae87-combined-ca-bundle\") pod \"barbican-keystone-listener-67bf9f4b94-nk9h8\" (UID: \"ae9a65ac-f1a7-4ef0-aeba-aca57985ae87\") " pod="openstack/barbican-keystone-listener-67bf9f4b94-nk9h8" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.360317 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxf9p\" (UniqueName: \"kubernetes.io/projected/ae9a65ac-f1a7-4ef0-aeba-aca57985ae87-kube-api-access-jxf9p\") pod \"barbican-keystone-listener-67bf9f4b94-nk9h8\" (UID: \"ae9a65ac-f1a7-4ef0-aeba-aca57985ae87\") " pod="openstack/barbican-keystone-listener-67bf9f4b94-nk9h8" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.369434 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-668f8fbd45-fpqbx" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.424653 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1650244d-222b-4c4e-80f8-720393e7eb67-logs\") pod \"barbican-api-59645bff48-kxnsb\" (UID: \"1650244d-222b-4c4e-80f8-720393e7eb67\") " pod="openstack/barbican-api-59645bff48-kxnsb" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.424697 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9947cc2-023b-407f-a6c2-b4b2e2c67455-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-gzfst\" (UID: \"b9947cc2-023b-407f-a6c2-b4b2e2c67455\") " pod="openstack/dnsmasq-dns-75c8ddd69c-gzfst" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.424714 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1650244d-222b-4c4e-80f8-720393e7eb67-config-data\") pod \"barbican-api-59645bff48-kxnsb\" (UID: \"1650244d-222b-4c4e-80f8-720393e7eb67\") " pod="openstack/barbican-api-59645bff48-kxnsb" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.424748 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1650244d-222b-4c4e-80f8-720393e7eb67-config-data-custom\") pod \"barbican-api-59645bff48-kxnsb\" (UID: \"1650244d-222b-4c4e-80f8-720393e7eb67\") " pod="openstack/barbican-api-59645bff48-kxnsb" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.424775 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4t4n\" (UniqueName: \"kubernetes.io/projected/b9947cc2-023b-407f-a6c2-b4b2e2c67455-kube-api-access-x4t4n\") pod \"dnsmasq-dns-75c8ddd69c-gzfst\" (UID: \"b9947cc2-023b-407f-a6c2-b4b2e2c67455\") " pod="openstack/dnsmasq-dns-75c8ddd69c-gzfst" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.424813 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b9947cc2-023b-407f-a6c2-b4b2e2c67455-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-gzfst\" (UID: \"b9947cc2-023b-407f-a6c2-b4b2e2c67455\") " pod="openstack/dnsmasq-dns-75c8ddd69c-gzfst" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.424849 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9947cc2-023b-407f-a6c2-b4b2e2c67455-config\") pod \"dnsmasq-dns-75c8ddd69c-gzfst\" (UID: \"b9947cc2-023b-407f-a6c2-b4b2e2c67455\") " pod="openstack/dnsmasq-dns-75c8ddd69c-gzfst" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.424907 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1650244d-222b-4c4e-80f8-720393e7eb67-combined-ca-bundle\") pod \"barbican-api-59645bff48-kxnsb\" (UID: \"1650244d-222b-4c4e-80f8-720393e7eb67\") " pod="openstack/barbican-api-59645bff48-kxnsb" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.424923 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4z2wj\" (UniqueName: \"kubernetes.io/projected/1650244d-222b-4c4e-80f8-720393e7eb67-kube-api-access-4z2wj\") pod \"barbican-api-59645bff48-kxnsb\" (UID: \"1650244d-222b-4c4e-80f8-720393e7eb67\") " pod="openstack/barbican-api-59645bff48-kxnsb" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.424984 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b9947cc2-023b-407f-a6c2-b4b2e2c67455-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-gzfst\" (UID: \"b9947cc2-023b-407f-a6c2-b4b2e2c67455\") " pod="openstack/dnsmasq-dns-75c8ddd69c-gzfst" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.425007 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b9947cc2-023b-407f-a6c2-b4b2e2c67455-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-gzfst\" (UID: \"b9947cc2-023b-407f-a6c2-b4b2e2c67455\") " pod="openstack/dnsmasq-dns-75c8ddd69c-gzfst" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.426408 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9947cc2-023b-407f-a6c2-b4b2e2c67455-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-gzfst\" (UID: \"b9947cc2-023b-407f-a6c2-b4b2e2c67455\") " pod="openstack/dnsmasq-dns-75c8ddd69c-gzfst" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.426627 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b9947cc2-023b-407f-a6c2-b4b2e2c67455-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-gzfst\" (UID: \"b9947cc2-023b-407f-a6c2-b4b2e2c67455\") " pod="openstack/dnsmasq-dns-75c8ddd69c-gzfst" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.427073 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-67bf9f4b94-nk9h8" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.428113 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b9947cc2-023b-407f-a6c2-b4b2e2c67455-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-gzfst\" (UID: \"b9947cc2-023b-407f-a6c2-b4b2e2c67455\") " pod="openstack/dnsmasq-dns-75c8ddd69c-gzfst" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.432524 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9947cc2-023b-407f-a6c2-b4b2e2c67455-config\") pod \"dnsmasq-dns-75c8ddd69c-gzfst\" (UID: \"b9947cc2-023b-407f-a6c2-b4b2e2c67455\") " pod="openstack/dnsmasq-dns-75c8ddd69c-gzfst" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.433410 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b9947cc2-023b-407f-a6c2-b4b2e2c67455-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-gzfst\" (UID: \"b9947cc2-023b-407f-a6c2-b4b2e2c67455\") " pod="openstack/dnsmasq-dns-75c8ddd69c-gzfst" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.447494 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4t4n\" (UniqueName: \"kubernetes.io/projected/b9947cc2-023b-407f-a6c2-b4b2e2c67455-kube-api-access-x4t4n\") pod \"dnsmasq-dns-75c8ddd69c-gzfst\" (UID: \"b9947cc2-023b-407f-a6c2-b4b2e2c67455\") " pod="openstack/dnsmasq-dns-75c8ddd69c-gzfst" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.526892 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1650244d-222b-4c4e-80f8-720393e7eb67-combined-ca-bundle\") pod \"barbican-api-59645bff48-kxnsb\" (UID: \"1650244d-222b-4c4e-80f8-720393e7eb67\") " pod="openstack/barbican-api-59645bff48-kxnsb" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.526947 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4z2wj\" (UniqueName: \"kubernetes.io/projected/1650244d-222b-4c4e-80f8-720393e7eb67-kube-api-access-4z2wj\") pod \"barbican-api-59645bff48-kxnsb\" (UID: \"1650244d-222b-4c4e-80f8-720393e7eb67\") " pod="openstack/barbican-api-59645bff48-kxnsb" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.527044 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1650244d-222b-4c4e-80f8-720393e7eb67-logs\") pod \"barbican-api-59645bff48-kxnsb\" (UID: \"1650244d-222b-4c4e-80f8-720393e7eb67\") " pod="openstack/barbican-api-59645bff48-kxnsb" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.527070 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1650244d-222b-4c4e-80f8-720393e7eb67-config-data\") pod \"barbican-api-59645bff48-kxnsb\" (UID: \"1650244d-222b-4c4e-80f8-720393e7eb67\") " pod="openstack/barbican-api-59645bff48-kxnsb" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.527105 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1650244d-222b-4c4e-80f8-720393e7eb67-config-data-custom\") pod \"barbican-api-59645bff48-kxnsb\" (UID: \"1650244d-222b-4c4e-80f8-720393e7eb67\") " pod="openstack/barbican-api-59645bff48-kxnsb" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.527782 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1650244d-222b-4c4e-80f8-720393e7eb67-logs\") pod \"barbican-api-59645bff48-kxnsb\" (UID: \"1650244d-222b-4c4e-80f8-720393e7eb67\") " pod="openstack/barbican-api-59645bff48-kxnsb" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.531864 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1650244d-222b-4c4e-80f8-720393e7eb67-config-data\") pod \"barbican-api-59645bff48-kxnsb\" (UID: \"1650244d-222b-4c4e-80f8-720393e7eb67\") " pod="openstack/barbican-api-59645bff48-kxnsb" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.534175 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1650244d-222b-4c4e-80f8-720393e7eb67-combined-ca-bundle\") pod \"barbican-api-59645bff48-kxnsb\" (UID: \"1650244d-222b-4c4e-80f8-720393e7eb67\") " pod="openstack/barbican-api-59645bff48-kxnsb" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.544245 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1650244d-222b-4c4e-80f8-720393e7eb67-config-data-custom\") pod \"barbican-api-59645bff48-kxnsb\" (UID: \"1650244d-222b-4c4e-80f8-720393e7eb67\") " pod="openstack/barbican-api-59645bff48-kxnsb" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.549040 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4z2wj\" (UniqueName: \"kubernetes.io/projected/1650244d-222b-4c4e-80f8-720393e7eb67-kube-api-access-4z2wj\") pod \"barbican-api-59645bff48-kxnsb\" (UID: \"1650244d-222b-4c4e-80f8-720393e7eb67\") " pod="openstack/barbican-api-59645bff48-kxnsb" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.595234 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5595d79c68-dgsth" podUID="5e9b87b2-4836-4857-8514-b2dcfe4de1f5" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.710447 4840 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.710484 4840 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.731538 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-94dfcc48d-mlrjd" podUID="1138b51b-edf9-42d2-8ea4-4d4b9122eae4" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.150:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.150:8443: connect: connection refused" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.741221 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-gzfst" Sep 30 14:13:28 crc kubenswrapper[4840]: I0930 14:13:28.760051 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-59645bff48-kxnsb" Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.305932 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-gxf58" Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.406855 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-9vd76" Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.422210 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.425773 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.445689 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b73e88ee-cb25-466d-9860-0dd54a1c1f4c-dns-swift-storage-0\") pod \"b73e88ee-cb25-466d-9860-0dd54a1c1f4c\" (UID: \"b73e88ee-cb25-466d-9860-0dd54a1c1f4c\") " Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.446101 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b73e88ee-cb25-466d-9860-0dd54a1c1f4c-ovsdbserver-nb\") pod \"b73e88ee-cb25-466d-9860-0dd54a1c1f4c\" (UID: \"b73e88ee-cb25-466d-9860-0dd54a1c1f4c\") " Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.446150 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lkflj\" (UniqueName: \"kubernetes.io/projected/b73e88ee-cb25-466d-9860-0dd54a1c1f4c-kube-api-access-lkflj\") pod \"b73e88ee-cb25-466d-9860-0dd54a1c1f4c\" (UID: \"b73e88ee-cb25-466d-9860-0dd54a1c1f4c\") " Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.446223 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b73e88ee-cb25-466d-9860-0dd54a1c1f4c-dns-svc\") pod \"b73e88ee-cb25-466d-9860-0dd54a1c1f4c\" (UID: \"b73e88ee-cb25-466d-9860-0dd54a1c1f4c\") " Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.446255 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b73e88ee-cb25-466d-9860-0dd54a1c1f4c-config\") pod \"b73e88ee-cb25-466d-9860-0dd54a1c1f4c\" (UID: \"b73e88ee-cb25-466d-9860-0dd54a1c1f4c\") " Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.446294 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b73e88ee-cb25-466d-9860-0dd54a1c1f4c-ovsdbserver-sb\") pod \"b73e88ee-cb25-466d-9860-0dd54a1c1f4c\" (UID: \"b73e88ee-cb25-466d-9860-0dd54a1c1f4c\") " Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.481980 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b73e88ee-cb25-466d-9860-0dd54a1c1f4c-kube-api-access-lkflj" (OuterVolumeSpecName: "kube-api-access-lkflj") pod "b73e88ee-cb25-466d-9860-0dd54a1c1f4c" (UID: "b73e88ee-cb25-466d-9860-0dd54a1c1f4c"). InnerVolumeSpecName "kube-api-access-lkflj". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.552067 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/242d13bb-6113-421b-83d1-fcccf91a9c68-db-sync-config-data\") pod \"242d13bb-6113-421b-83d1-fcccf91a9c68\" (UID: \"242d13bb-6113-421b-83d1-fcccf91a9c68\") " Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.552377 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/242d13bb-6113-421b-83d1-fcccf91a9c68-etc-machine-id\") pod \"242d13bb-6113-421b-83d1-fcccf91a9c68\" (UID: \"242d13bb-6113-421b-83d1-fcccf91a9c68\") " Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.552433 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjbfr\" (UniqueName: \"kubernetes.io/projected/242d13bb-6113-421b-83d1-fcccf91a9c68-kube-api-access-pjbfr\") pod \"242d13bb-6113-421b-83d1-fcccf91a9c68\" (UID: \"242d13bb-6113-421b-83d1-fcccf91a9c68\") " Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.552489 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/242d13bb-6113-421b-83d1-fcccf91a9c68-scripts\") pod \"242d13bb-6113-421b-83d1-fcccf91a9c68\" (UID: \"242d13bb-6113-421b-83d1-fcccf91a9c68\") " Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.552517 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/242d13bb-6113-421b-83d1-fcccf91a9c68-config-data\") pod \"242d13bb-6113-421b-83d1-fcccf91a9c68\" (UID: \"242d13bb-6113-421b-83d1-fcccf91a9c68\") " Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.552601 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/242d13bb-6113-421b-83d1-fcccf91a9c68-combined-ca-bundle\") pod \"242d13bb-6113-421b-83d1-fcccf91a9c68\" (UID: \"242d13bb-6113-421b-83d1-fcccf91a9c68\") " Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.553033 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lkflj\" (UniqueName: \"kubernetes.io/projected/b73e88ee-cb25-466d-9860-0dd54a1c1f4c-kube-api-access-lkflj\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.557602 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/242d13bb-6113-421b-83d1-fcccf91a9c68-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "242d13bb-6113-421b-83d1-fcccf91a9c68" (UID: "242d13bb-6113-421b-83d1-fcccf91a9c68"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.559846 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/242d13bb-6113-421b-83d1-fcccf91a9c68-kube-api-access-pjbfr" (OuterVolumeSpecName: "kube-api-access-pjbfr") pod "242d13bb-6113-421b-83d1-fcccf91a9c68" (UID: "242d13bb-6113-421b-83d1-fcccf91a9c68"). InnerVolumeSpecName "kube-api-access-pjbfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.563440 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b73e88ee-cb25-466d-9860-0dd54a1c1f4c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b73e88ee-cb25-466d-9860-0dd54a1c1f4c" (UID: "b73e88ee-cb25-466d-9860-0dd54a1c1f4c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.565761 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/242d13bb-6113-421b-83d1-fcccf91a9c68-scripts" (OuterVolumeSpecName: "scripts") pod "242d13bb-6113-421b-83d1-fcccf91a9c68" (UID: "242d13bb-6113-421b-83d1-fcccf91a9c68"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.582152 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/242d13bb-6113-421b-83d1-fcccf91a9c68-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "242d13bb-6113-421b-83d1-fcccf91a9c68" (UID: "242d13bb-6113-421b-83d1-fcccf91a9c68"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.640150 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b73e88ee-cb25-466d-9860-0dd54a1c1f4c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b73e88ee-cb25-466d-9860-0dd54a1c1f4c" (UID: "b73e88ee-cb25-466d-9860-0dd54a1c1f4c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.654127 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/242d13bb-6113-421b-83d1-fcccf91a9c68-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "242d13bb-6113-421b-83d1-fcccf91a9c68" (UID: "242d13bb-6113-421b-83d1-fcccf91a9c68"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.655340 4840 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/242d13bb-6113-421b-83d1-fcccf91a9c68-etc-machine-id\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.655414 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjbfr\" (UniqueName: \"kubernetes.io/projected/242d13bb-6113-421b-83d1-fcccf91a9c68-kube-api-access-pjbfr\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.655473 4840 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/242d13bb-6113-421b-83d1-fcccf91a9c68-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.655526 4840 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b73e88ee-cb25-466d-9860-0dd54a1c1f4c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.659734 4840 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b73e88ee-cb25-466d-9860-0dd54a1c1f4c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.659839 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/242d13bb-6113-421b-83d1-fcccf91a9c68-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.659917 4840 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/242d13bb-6113-421b-83d1-fcccf91a9c68-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.706610 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b73e88ee-cb25-466d-9860-0dd54a1c1f4c-config" (OuterVolumeSpecName: "config") pod "b73e88ee-cb25-466d-9860-0dd54a1c1f4c" (UID: "b73e88ee-cb25-466d-9860-0dd54a1c1f4c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.716715 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b73e88ee-cb25-466d-9860-0dd54a1c1f4c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b73e88ee-cb25-466d-9860-0dd54a1c1f4c" (UID: "b73e88ee-cb25-466d-9860-0dd54a1c1f4c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.723534 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-gxf58" event={"ID":"b73e88ee-cb25-466d-9860-0dd54a1c1f4c","Type":"ContainerDied","Data":"b2c37df544a67982abf51985e156af05ad950bcc25628ce8bd5c54117c6d35eb"} Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.723618 4840 scope.go:117] "RemoveContainer" containerID="8f1dc45f77092a5778340c020c0952b71db3f7fad8ff85b132ad5c8cc6e7a5c5" Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.723734 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-gxf58" Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.726598 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-9vd76" Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.727656 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-9vd76" event={"ID":"242d13bb-6113-421b-83d1-fcccf91a9c68","Type":"ContainerDied","Data":"e7ae2c59a20eb580f9d27eb067fb6d3b0b51bff398919efd4d34d75f936615d9"} Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.727677 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e7ae2c59a20eb580f9d27eb067fb6d3b0b51bff398919efd4d34d75f936615d9" Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.727736 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/242d13bb-6113-421b-83d1-fcccf91a9c68-config-data" (OuterVolumeSpecName: "config-data") pod "242d13bb-6113-421b-83d1-fcccf91a9c68" (UID: "242d13bb-6113-421b-83d1-fcccf91a9c68"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.736154 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b73e88ee-cb25-466d-9860-0dd54a1c1f4c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b73e88ee-cb25-466d-9860-0dd54a1c1f4c" (UID: "b73e88ee-cb25-466d-9860-0dd54a1c1f4c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.769513 4840 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b73e88ee-cb25-466d-9860-0dd54a1c1f4c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.769817 4840 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b73e88ee-cb25-466d-9860-0dd54a1c1f4c-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.769826 4840 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b73e88ee-cb25-466d-9860-0dd54a1c1f4c-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.769835 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/242d13bb-6113-421b-83d1-fcccf91a9c68-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:29 crc kubenswrapper[4840]: I0930 14:13:29.775798 4840 scope.go:117] "RemoveContainer" containerID="67c24bcbe6a8e03ad93ea8400fa9337d59ee9b410dc1bf72342e0e16d1c8942a" Sep 30 14:13:29 crc kubenswrapper[4840]: E0930 14:13:29.865320 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="e4a730c4-623d-4ced-a687-bec2d627eae9" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.289730 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-gxf58"] Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.380612 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-gxf58"] Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.420516 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-668f8fbd45-fpqbx"] Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.438801 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-gzfst"] Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.465488 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-67bf9f4b94-nk9h8"] Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.476026 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-59645bff48-kxnsb"] Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.484534 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6cbd5fdcfb-hn4pn"] Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.696260 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Sep 30 14:13:30 crc kubenswrapper[4840]: E0930 14:13:30.701122 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b73e88ee-cb25-466d-9860-0dd54a1c1f4c" containerName="dnsmasq-dns" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.701147 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="b73e88ee-cb25-466d-9860-0dd54a1c1f4c" containerName="dnsmasq-dns" Sep 30 14:13:30 crc kubenswrapper[4840]: E0930 14:13:30.701161 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="242d13bb-6113-421b-83d1-fcccf91a9c68" containerName="cinder-db-sync" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.701167 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="242d13bb-6113-421b-83d1-fcccf91a9c68" containerName="cinder-db-sync" Sep 30 14:13:30 crc kubenswrapper[4840]: E0930 14:13:30.701197 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b73e88ee-cb25-466d-9860-0dd54a1c1f4c" containerName="init" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.701203 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="b73e88ee-cb25-466d-9860-0dd54a1c1f4c" containerName="init" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.701357 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="242d13bb-6113-421b-83d1-fcccf91a9c68" containerName="cinder-db-sync" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.701376 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="b73e88ee-cb25-466d-9860-0dd54a1c1f4c" containerName="dnsmasq-dns" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.702383 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.713755 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.713933 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-2qzkw" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.714049 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.714212 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.714387 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.763118 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-gzfst"] Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.772289 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-59645bff48-kxnsb" event={"ID":"1650244d-222b-4c4e-80f8-720393e7eb67","Type":"ContainerStarted","Data":"97be88e5e433c272931edd0cb1c41d8d463cffabc5ac517bbe7ea19f723986a9"} Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.795226 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6cbd5fdcfb-hn4pn" event={"ID":"c1eaca45-6337-4f35-ac67-dfa0fc83e1bb","Type":"ContainerStarted","Data":"110ae8ccd0fe24a3bf0bd23f65281f2d4a24f8656495b4a6025cc87d2169c405"} Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.798846 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-qs9s6"] Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.801409 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-668f8fbd45-fpqbx" event={"ID":"0fa943f1-8698-41e3-a893-fd1c8845076a","Type":"ContainerStarted","Data":"6387bbd4a2e255c47b01b4798a5bece6760939874209544ba2a6746624c9c67c"} Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.801493 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-qs9s6" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.821439 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-qs9s6"] Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.826614 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3\") " pod="openstack/cinder-scheduler-0" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.826691 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3\") " pod="openstack/cinder-scheduler-0" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.826752 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3\") " pod="openstack/cinder-scheduler-0" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.826776 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qrqw\" (UniqueName: \"kubernetes.io/projected/9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3-kube-api-access-2qrqw\") pod \"cinder-scheduler-0\" (UID: \"9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3\") " pod="openstack/cinder-scheduler-0" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.826799 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3-config-data\") pod \"cinder-scheduler-0\" (UID: \"9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3\") " pod="openstack/cinder-scheduler-0" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.826825 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3-scripts\") pod \"cinder-scheduler-0\" (UID: \"9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3\") " pod="openstack/cinder-scheduler-0" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.829772 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-67bf9f4b94-nk9h8" event={"ID":"ae9a65ac-f1a7-4ef0-aeba-aca57985ae87","Type":"ContainerStarted","Data":"bff295d4ad9f66ad5e5971c8814d1e9478d5b5a19c41d081f6d9086cd203907b"} Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.835629 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-gzfst" event={"ID":"b9947cc2-023b-407f-a6c2-b4b2e2c67455","Type":"ContainerStarted","Data":"ef19eb81aae7be13f28158ed323aadac2b95173b51af1609bc82cbd72ba7ee4a"} Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.835831 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-75c8ddd69c-gzfst" podUID="b9947cc2-023b-407f-a6c2-b4b2e2c67455" containerName="init" containerID="cri-o://ffedbc29b530841a9aef3435a7c7dff76e3b6de62a90dbb186ccdb322d1c5b50" gracePeriod=10 Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.856179 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e4a730c4-623d-4ced-a687-bec2d627eae9","Type":"ContainerStarted","Data":"8f4c2238d84a252e99a591f48de599db61ae31531f66f05d1cefccc93adf1543"} Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.856543 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e4a730c4-623d-4ced-a687-bec2d627eae9" containerName="ceilometer-notification-agent" containerID="cri-o://19c1bb2cbbc34614446d73b508c35aafb815b7a569fbfdb49916e9cdaadb3519" gracePeriod=30 Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.856634 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.856651 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e4a730c4-623d-4ced-a687-bec2d627eae9" containerName="proxy-httpd" containerID="cri-o://8f4c2238d84a252e99a591f48de599db61ae31531f66f05d1cefccc93adf1543" gracePeriod=30 Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.856685 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e4a730c4-623d-4ced-a687-bec2d627eae9" containerName="sg-core" containerID="cri-o://f2ce0409cad8baf6e2ed54568f34ac02c6b59d18905648cee072364d2def57b6" gracePeriod=30 Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.932217 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3\") " pod="openstack/cinder-scheduler-0" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.932297 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae2f3329-33ca-402d-b841-8f1e257bbf8a-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-qs9s6\" (UID: \"ae2f3329-33ca-402d-b841-8f1e257bbf8a\") " pod="openstack/dnsmasq-dns-5784cf869f-qs9s6" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.932381 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3\") " pod="openstack/cinder-scheduler-0" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.932496 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3\") " pod="openstack/cinder-scheduler-0" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.932523 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qrqw\" (UniqueName: \"kubernetes.io/projected/9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3-kube-api-access-2qrqw\") pod \"cinder-scheduler-0\" (UID: \"9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3\") " pod="openstack/cinder-scheduler-0" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.932591 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3-config-data\") pod \"cinder-scheduler-0\" (UID: \"9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3\") " pod="openstack/cinder-scheduler-0" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.932642 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae2f3329-33ca-402d-b841-8f1e257bbf8a-dns-svc\") pod \"dnsmasq-dns-5784cf869f-qs9s6\" (UID: \"ae2f3329-33ca-402d-b841-8f1e257bbf8a\") " pod="openstack/dnsmasq-dns-5784cf869f-qs9s6" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.932657 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3-scripts\") pod \"cinder-scheduler-0\" (UID: \"9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3\") " pod="openstack/cinder-scheduler-0" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.932681 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ae2f3329-33ca-402d-b841-8f1e257bbf8a-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-qs9s6\" (UID: \"ae2f3329-33ca-402d-b841-8f1e257bbf8a\") " pod="openstack/dnsmasq-dns-5784cf869f-qs9s6" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.932723 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae2f3329-33ca-402d-b841-8f1e257bbf8a-config\") pod \"dnsmasq-dns-5784cf869f-qs9s6\" (UID: \"ae2f3329-33ca-402d-b841-8f1e257bbf8a\") " pod="openstack/dnsmasq-dns-5784cf869f-qs9s6" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.932742 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrvf9\" (UniqueName: \"kubernetes.io/projected/ae2f3329-33ca-402d-b841-8f1e257bbf8a-kube-api-access-nrvf9\") pod \"dnsmasq-dns-5784cf869f-qs9s6\" (UID: \"ae2f3329-33ca-402d-b841-8f1e257bbf8a\") " pod="openstack/dnsmasq-dns-5784cf869f-qs9s6" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.932796 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ae2f3329-33ca-402d-b841-8f1e257bbf8a-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-qs9s6\" (UID: \"ae2f3329-33ca-402d-b841-8f1e257bbf8a\") " pod="openstack/dnsmasq-dns-5784cf869f-qs9s6" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.934835 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3\") " pod="openstack/cinder-scheduler-0" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.941230 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3\") " pod="openstack/cinder-scheduler-0" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.944951 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3-config-data\") pod \"cinder-scheduler-0\" (UID: \"9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3\") " pod="openstack/cinder-scheduler-0" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.946812 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3-scripts\") pod \"cinder-scheduler-0\" (UID: \"9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3\") " pod="openstack/cinder-scheduler-0" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.947237 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3\") " pod="openstack/cinder-scheduler-0" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.961270 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.963139 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.965671 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.965989 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qrqw\" (UniqueName: \"kubernetes.io/projected/9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3-kube-api-access-2qrqw\") pod \"cinder-scheduler-0\" (UID: \"9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3\") " pod="openstack/cinder-scheduler-0" Sep 30 14:13:30 crc kubenswrapper[4840]: I0930 14:13:30.976606 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.034514 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae2f3329-33ca-402d-b841-8f1e257bbf8a-dns-svc\") pod \"dnsmasq-dns-5784cf869f-qs9s6\" (UID: \"ae2f3329-33ca-402d-b841-8f1e257bbf8a\") " pod="openstack/dnsmasq-dns-5784cf869f-qs9s6" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.036149 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ae2f3329-33ca-402d-b841-8f1e257bbf8a-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-qs9s6\" (UID: \"ae2f3329-33ca-402d-b841-8f1e257bbf8a\") " pod="openstack/dnsmasq-dns-5784cf869f-qs9s6" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.036207 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae2f3329-33ca-402d-b841-8f1e257bbf8a-config\") pod \"dnsmasq-dns-5784cf869f-qs9s6\" (UID: \"ae2f3329-33ca-402d-b841-8f1e257bbf8a\") " pod="openstack/dnsmasq-dns-5784cf869f-qs9s6" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.036238 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrvf9\" (UniqueName: \"kubernetes.io/projected/ae2f3329-33ca-402d-b841-8f1e257bbf8a-kube-api-access-nrvf9\") pod \"dnsmasq-dns-5784cf869f-qs9s6\" (UID: \"ae2f3329-33ca-402d-b841-8f1e257bbf8a\") " pod="openstack/dnsmasq-dns-5784cf869f-qs9s6" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.036320 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ae2f3329-33ca-402d-b841-8f1e257bbf8a-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-qs9s6\" (UID: \"ae2f3329-33ca-402d-b841-8f1e257bbf8a\") " pod="openstack/dnsmasq-dns-5784cf869f-qs9s6" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.036771 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae2f3329-33ca-402d-b841-8f1e257bbf8a-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-qs9s6\" (UID: \"ae2f3329-33ca-402d-b841-8f1e257bbf8a\") " pod="openstack/dnsmasq-dns-5784cf869f-qs9s6" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.036054 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae2f3329-33ca-402d-b841-8f1e257bbf8a-dns-svc\") pod \"dnsmasq-dns-5784cf869f-qs9s6\" (UID: \"ae2f3329-33ca-402d-b841-8f1e257bbf8a\") " pod="openstack/dnsmasq-dns-5784cf869f-qs9s6" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.037771 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae2f3329-33ca-402d-b841-8f1e257bbf8a-config\") pod \"dnsmasq-dns-5784cf869f-qs9s6\" (UID: \"ae2f3329-33ca-402d-b841-8f1e257bbf8a\") " pod="openstack/dnsmasq-dns-5784cf869f-qs9s6" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.038529 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ae2f3329-33ca-402d-b841-8f1e257bbf8a-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-qs9s6\" (UID: \"ae2f3329-33ca-402d-b841-8f1e257bbf8a\") " pod="openstack/dnsmasq-dns-5784cf869f-qs9s6" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.038933 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ae2f3329-33ca-402d-b841-8f1e257bbf8a-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-qs9s6\" (UID: \"ae2f3329-33ca-402d-b841-8f1e257bbf8a\") " pod="openstack/dnsmasq-dns-5784cf869f-qs9s6" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.039484 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae2f3329-33ca-402d-b841-8f1e257bbf8a-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-qs9s6\" (UID: \"ae2f3329-33ca-402d-b841-8f1e257bbf8a\") " pod="openstack/dnsmasq-dns-5784cf869f-qs9s6" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.058598 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrvf9\" (UniqueName: \"kubernetes.io/projected/ae2f3329-33ca-402d-b841-8f1e257bbf8a-kube-api-access-nrvf9\") pod \"dnsmasq-dns-5784cf869f-qs9s6\" (UID: \"ae2f3329-33ca-402d-b841-8f1e257bbf8a\") " pod="openstack/dnsmasq-dns-5784cf869f-qs9s6" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.073043 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.138928 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22734746-33bf-495a-9c55-b7c4cf274fa7-config-data\") pod \"cinder-api-0\" (UID: \"22734746-33bf-495a-9c55-b7c4cf274fa7\") " pod="openstack/cinder-api-0" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.139064 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22734746-33bf-495a-9c55-b7c4cf274fa7-scripts\") pod \"cinder-api-0\" (UID: \"22734746-33bf-495a-9c55-b7c4cf274fa7\") " pod="openstack/cinder-api-0" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.139148 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/22734746-33bf-495a-9c55-b7c4cf274fa7-config-data-custom\") pod \"cinder-api-0\" (UID: \"22734746-33bf-495a-9c55-b7c4cf274fa7\") " pod="openstack/cinder-api-0" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.139289 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/22734746-33bf-495a-9c55-b7c4cf274fa7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"22734746-33bf-495a-9c55-b7c4cf274fa7\") " pod="openstack/cinder-api-0" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.139313 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22734746-33bf-495a-9c55-b7c4cf274fa7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"22734746-33bf-495a-9c55-b7c4cf274fa7\") " pod="openstack/cinder-api-0" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.139369 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzxfv\" (UniqueName: \"kubernetes.io/projected/22734746-33bf-495a-9c55-b7c4cf274fa7-kube-api-access-xzxfv\") pod \"cinder-api-0\" (UID: \"22734746-33bf-495a-9c55-b7c4cf274fa7\") " pod="openstack/cinder-api-0" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.139402 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22734746-33bf-495a-9c55-b7c4cf274fa7-logs\") pod \"cinder-api-0\" (UID: \"22734746-33bf-495a-9c55-b7c4cf274fa7\") " pod="openstack/cinder-api-0" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.183676 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-qs9s6" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.240897 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/22734746-33bf-495a-9c55-b7c4cf274fa7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"22734746-33bf-495a-9c55-b7c4cf274fa7\") " pod="openstack/cinder-api-0" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.240943 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22734746-33bf-495a-9c55-b7c4cf274fa7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"22734746-33bf-495a-9c55-b7c4cf274fa7\") " pod="openstack/cinder-api-0" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.240963 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzxfv\" (UniqueName: \"kubernetes.io/projected/22734746-33bf-495a-9c55-b7c4cf274fa7-kube-api-access-xzxfv\") pod \"cinder-api-0\" (UID: \"22734746-33bf-495a-9c55-b7c4cf274fa7\") " pod="openstack/cinder-api-0" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.240983 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22734746-33bf-495a-9c55-b7c4cf274fa7-logs\") pod \"cinder-api-0\" (UID: \"22734746-33bf-495a-9c55-b7c4cf274fa7\") " pod="openstack/cinder-api-0" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.241051 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22734746-33bf-495a-9c55-b7c4cf274fa7-config-data\") pod \"cinder-api-0\" (UID: \"22734746-33bf-495a-9c55-b7c4cf274fa7\") " pod="openstack/cinder-api-0" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.241076 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22734746-33bf-495a-9c55-b7c4cf274fa7-scripts\") pod \"cinder-api-0\" (UID: \"22734746-33bf-495a-9c55-b7c4cf274fa7\") " pod="openstack/cinder-api-0" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.241121 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/22734746-33bf-495a-9c55-b7c4cf274fa7-config-data-custom\") pod \"cinder-api-0\" (UID: \"22734746-33bf-495a-9c55-b7c4cf274fa7\") " pod="openstack/cinder-api-0" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.246420 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22734746-33bf-495a-9c55-b7c4cf274fa7-logs\") pod \"cinder-api-0\" (UID: \"22734746-33bf-495a-9c55-b7c4cf274fa7\") " pod="openstack/cinder-api-0" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.246500 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/22734746-33bf-495a-9c55-b7c4cf274fa7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"22734746-33bf-495a-9c55-b7c4cf274fa7\") " pod="openstack/cinder-api-0" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.276144 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22734746-33bf-495a-9c55-b7c4cf274fa7-scripts\") pod \"cinder-api-0\" (UID: \"22734746-33bf-495a-9c55-b7c4cf274fa7\") " pod="openstack/cinder-api-0" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.281169 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22734746-33bf-495a-9c55-b7c4cf274fa7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"22734746-33bf-495a-9c55-b7c4cf274fa7\") " pod="openstack/cinder-api-0" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.285714 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22734746-33bf-495a-9c55-b7c4cf274fa7-config-data\") pod \"cinder-api-0\" (UID: \"22734746-33bf-495a-9c55-b7c4cf274fa7\") " pod="openstack/cinder-api-0" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.286191 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/22734746-33bf-495a-9c55-b7c4cf274fa7-config-data-custom\") pod \"cinder-api-0\" (UID: \"22734746-33bf-495a-9c55-b7c4cf274fa7\") " pod="openstack/cinder-api-0" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.326760 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzxfv\" (UniqueName: \"kubernetes.io/projected/22734746-33bf-495a-9c55-b7c4cf274fa7-kube-api-access-xzxfv\") pod \"cinder-api-0\" (UID: \"22734746-33bf-495a-9c55-b7c4cf274fa7\") " pod="openstack/cinder-api-0" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.362984 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.521008 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-gzfst" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.652092 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9947cc2-023b-407f-a6c2-b4b2e2c67455-dns-svc\") pod \"b9947cc2-023b-407f-a6c2-b4b2e2c67455\" (UID: \"b9947cc2-023b-407f-a6c2-b4b2e2c67455\") " Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.652171 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b9947cc2-023b-407f-a6c2-b4b2e2c67455-dns-swift-storage-0\") pod \"b9947cc2-023b-407f-a6c2-b4b2e2c67455\" (UID: \"b9947cc2-023b-407f-a6c2-b4b2e2c67455\") " Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.652326 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4t4n\" (UniqueName: \"kubernetes.io/projected/b9947cc2-023b-407f-a6c2-b4b2e2c67455-kube-api-access-x4t4n\") pod \"b9947cc2-023b-407f-a6c2-b4b2e2c67455\" (UID: \"b9947cc2-023b-407f-a6c2-b4b2e2c67455\") " Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.652355 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9947cc2-023b-407f-a6c2-b4b2e2c67455-config\") pod \"b9947cc2-023b-407f-a6c2-b4b2e2c67455\" (UID: \"b9947cc2-023b-407f-a6c2-b4b2e2c67455\") " Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.652478 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b9947cc2-023b-407f-a6c2-b4b2e2c67455-ovsdbserver-nb\") pod \"b9947cc2-023b-407f-a6c2-b4b2e2c67455\" (UID: \"b9947cc2-023b-407f-a6c2-b4b2e2c67455\") " Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.652500 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b9947cc2-023b-407f-a6c2-b4b2e2c67455-ovsdbserver-sb\") pod \"b9947cc2-023b-407f-a6c2-b4b2e2c67455\" (UID: \"b9947cc2-023b-407f-a6c2-b4b2e2c67455\") " Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.663391 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9947cc2-023b-407f-a6c2-b4b2e2c67455-kube-api-access-x4t4n" (OuterVolumeSpecName: "kube-api-access-x4t4n") pod "b9947cc2-023b-407f-a6c2-b4b2e2c67455" (UID: "b9947cc2-023b-407f-a6c2-b4b2e2c67455"). InnerVolumeSpecName "kube-api-access-x4t4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.689449 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9947cc2-023b-407f-a6c2-b4b2e2c67455-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b9947cc2-023b-407f-a6c2-b4b2e2c67455" (UID: "b9947cc2-023b-407f-a6c2-b4b2e2c67455"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.690561 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9947cc2-023b-407f-a6c2-b4b2e2c67455-config" (OuterVolumeSpecName: "config") pod "b9947cc2-023b-407f-a6c2-b4b2e2c67455" (UID: "b9947cc2-023b-407f-a6c2-b4b2e2c67455"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.699807 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9947cc2-023b-407f-a6c2-b4b2e2c67455-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b9947cc2-023b-407f-a6c2-b4b2e2c67455" (UID: "b9947cc2-023b-407f-a6c2-b4b2e2c67455"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.714474 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9947cc2-023b-407f-a6c2-b4b2e2c67455-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b9947cc2-023b-407f-a6c2-b4b2e2c67455" (UID: "b9947cc2-023b-407f-a6c2-b4b2e2c67455"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.757617 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9947cc2-023b-407f-a6c2-b4b2e2c67455-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b9947cc2-023b-407f-a6c2-b4b2e2c67455" (UID: "b9947cc2-023b-407f-a6c2-b4b2e2c67455"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.757754 4840 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b9947cc2-023b-407f-a6c2-b4b2e2c67455-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.757771 4840 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b9947cc2-023b-407f-a6c2-b4b2e2c67455-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.757784 4840 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b9947cc2-023b-407f-a6c2-b4b2e2c67455-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.757795 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4t4n\" (UniqueName: \"kubernetes.io/projected/b9947cc2-023b-407f-a6c2-b4b2e2c67455-kube-api-access-x4t4n\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.757808 4840 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9947cc2-023b-407f-a6c2-b4b2e2c67455-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.856881 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.868086 4840 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9947cc2-023b-407f-a6c2-b4b2e2c67455-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.909626 4840 generic.go:334] "Generic (PLEG): container finished" podID="e4a730c4-623d-4ced-a687-bec2d627eae9" containerID="8f4c2238d84a252e99a591f48de599db61ae31531f66f05d1cefccc93adf1543" exitCode=0 Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.909655 4840 generic.go:334] "Generic (PLEG): container finished" podID="e4a730c4-623d-4ced-a687-bec2d627eae9" containerID="f2ce0409cad8baf6e2ed54568f34ac02c6b59d18905648cee072364d2def57b6" exitCode=2 Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.909698 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e4a730c4-623d-4ced-a687-bec2d627eae9","Type":"ContainerDied","Data":"8f4c2238d84a252e99a591f48de599db61ae31531f66f05d1cefccc93adf1543"} Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.909723 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e4a730c4-623d-4ced-a687-bec2d627eae9","Type":"ContainerDied","Data":"f2ce0409cad8baf6e2ed54568f34ac02c6b59d18905648cee072364d2def57b6"} Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.929184 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-59645bff48-kxnsb" event={"ID":"1650244d-222b-4c4e-80f8-720393e7eb67","Type":"ContainerStarted","Data":"9bbeaf4ecaa27d44f6cc27c9d44e044a6e6f58d68591aa17c80309694e88e34e"} Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.929244 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-59645bff48-kxnsb" event={"ID":"1650244d-222b-4c4e-80f8-720393e7eb67","Type":"ContainerStarted","Data":"af005a4cb3f5c14bbf6ea4d65b23bee0a91d02cf2ba5b71bf0fd3c936f7904a4"} Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.929279 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-59645bff48-kxnsb" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.929326 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-59645bff48-kxnsb" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.936711 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3","Type":"ContainerStarted","Data":"861cc27309a90677d4bc9cd4e3e49f491810252bcd526bcdb2faea77f3b9fea5"} Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.961068 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6cbd5fdcfb-hn4pn" event={"ID":"c1eaca45-6337-4f35-ac67-dfa0fc83e1bb","Type":"ContainerStarted","Data":"8ddd75908b80990160de3c869630c4e6b1d291417543472dbe84e88756b3d662"} Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.961151 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6cbd5fdcfb-hn4pn" event={"ID":"c1eaca45-6337-4f35-ac67-dfa0fc83e1bb","Type":"ContainerStarted","Data":"b8ae58557b2f320938af74209e68edb0bab722966e1d3f922888ab6fa88dd7be"} Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.961211 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6cbd5fdcfb-hn4pn" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.961241 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6cbd5fdcfb-hn4pn" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.965924 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-59645bff48-kxnsb" podStartSLOduration=3.9659073080000002 podStartE2EDuration="3.965907308s" podCreationTimestamp="2025-09-30 14:13:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:13:31.961205745 +0000 UTC m=+1040.590292168" watchObservedRunningTime="2025-09-30 14:13:31.965907308 +0000 UTC m=+1040.594993731" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.967897 4840 generic.go:334] "Generic (PLEG): container finished" podID="b9947cc2-023b-407f-a6c2-b4b2e2c67455" containerID="ffedbc29b530841a9aef3435a7c7dff76e3b6de62a90dbb186ccdb322d1c5b50" exitCode=0 Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.967946 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-gzfst" event={"ID":"b9947cc2-023b-407f-a6c2-b4b2e2c67455","Type":"ContainerDied","Data":"ef19eb81aae7be13f28158ed323aadac2b95173b51af1609bc82cbd72ba7ee4a"} Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.967975 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-gzfst" event={"ID":"b9947cc2-023b-407f-a6c2-b4b2e2c67455","Type":"ContainerDied","Data":"ffedbc29b530841a9aef3435a7c7dff76e3b6de62a90dbb186ccdb322d1c5b50"} Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.967994 4840 scope.go:117] "RemoveContainer" containerID="ffedbc29b530841a9aef3435a7c7dff76e3b6de62a90dbb186ccdb322d1c5b50" Sep 30 14:13:31 crc kubenswrapper[4840]: I0930 14:13:31.968158 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-gzfst" Sep 30 14:13:32 crc kubenswrapper[4840]: I0930 14:13:32.044706 4840 scope.go:117] "RemoveContainer" containerID="ffedbc29b530841a9aef3435a7c7dff76e3b6de62a90dbb186ccdb322d1c5b50" Sep 30 14:13:32 crc kubenswrapper[4840]: E0930 14:13:32.046368 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffedbc29b530841a9aef3435a7c7dff76e3b6de62a90dbb186ccdb322d1c5b50\": container with ID starting with ffedbc29b530841a9aef3435a7c7dff76e3b6de62a90dbb186ccdb322d1c5b50 not found: ID does not exist" containerID="ffedbc29b530841a9aef3435a7c7dff76e3b6de62a90dbb186ccdb322d1c5b50" Sep 30 14:13:32 crc kubenswrapper[4840]: I0930 14:13:32.046407 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffedbc29b530841a9aef3435a7c7dff76e3b6de62a90dbb186ccdb322d1c5b50"} err="failed to get container status \"ffedbc29b530841a9aef3435a7c7dff76e3b6de62a90dbb186ccdb322d1c5b50\": rpc error: code = NotFound desc = could not find container \"ffedbc29b530841a9aef3435a7c7dff76e3b6de62a90dbb186ccdb322d1c5b50\": container with ID starting with ffedbc29b530841a9aef3435a7c7dff76e3b6de62a90dbb186ccdb322d1c5b50 not found: ID does not exist" Sep 30 14:13:32 crc kubenswrapper[4840]: I0930 14:13:32.112443 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-6cbd5fdcfb-hn4pn" podStartSLOduration=5.112424729 podStartE2EDuration="5.112424729s" podCreationTimestamp="2025-09-30 14:13:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:13:31.984652357 +0000 UTC m=+1040.613738780" watchObservedRunningTime="2025-09-30 14:13:32.112424729 +0000 UTC m=+1040.741511152" Sep 30 14:13:32 crc kubenswrapper[4840]: I0930 14:13:32.161704 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b73e88ee-cb25-466d-9860-0dd54a1c1f4c" path="/var/lib/kubelet/pods/b73e88ee-cb25-466d-9860-0dd54a1c1f4c/volumes" Sep 30 14:13:32 crc kubenswrapper[4840]: I0930 14:13:32.162305 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-qs9s6"] Sep 30 14:13:32 crc kubenswrapper[4840]: I0930 14:13:32.162331 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Sep 30 14:13:32 crc kubenswrapper[4840]: I0930 14:13:32.162345 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-gzfst"] Sep 30 14:13:32 crc kubenswrapper[4840]: I0930 14:13:32.171252 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-gzfst"] Sep 30 14:13:32 crc kubenswrapper[4840]: I0930 14:13:32.980108 4840 generic.go:334] "Generic (PLEG): container finished" podID="ae2f3329-33ca-402d-b841-8f1e257bbf8a" containerID="218b0c923a2f1127faf075a7b6fae3b7c0e6693b90820678f8522b3ef476766a" exitCode=0 Sep 30 14:13:32 crc kubenswrapper[4840]: I0930 14:13:32.980216 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-qs9s6" event={"ID":"ae2f3329-33ca-402d-b841-8f1e257bbf8a","Type":"ContainerDied","Data":"218b0c923a2f1127faf075a7b6fae3b7c0e6693b90820678f8522b3ef476766a"} Sep 30 14:13:32 crc kubenswrapper[4840]: I0930 14:13:32.980535 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-qs9s6" event={"ID":"ae2f3329-33ca-402d-b841-8f1e257bbf8a","Type":"ContainerStarted","Data":"6789707326d38b24e2609c00bdd94a43191af819d0e1ad323c450ef1f245470b"} Sep 30 14:13:32 crc kubenswrapper[4840]: I0930 14:13:32.987950 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"22734746-33bf-495a-9c55-b7c4cf274fa7","Type":"ContainerStarted","Data":"80927d30a386a824baebf38a057bc97ccb64ade654bbc94b57438f9bbafee003"} Sep 30 14:13:32 crc kubenswrapper[4840]: I0930 14:13:32.987999 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"22734746-33bf-495a-9c55-b7c4cf274fa7","Type":"ContainerStarted","Data":"fb02f64294175d668c0e3e208d67842b7558e1f80e91f393ed5924c5a26ee556"} Sep 30 14:13:34 crc kubenswrapper[4840]: I0930 14:13:34.128283 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9947cc2-023b-407f-a6c2-b4b2e2c67455" path="/var/lib/kubelet/pods/b9947cc2-023b-407f-a6c2-b4b2e2c67455/volumes" Sep 30 14:13:34 crc kubenswrapper[4840]: I0930 14:13:34.777197 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.018497 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-668f8fbd45-fpqbx" event={"ID":"0fa943f1-8698-41e3-a893-fd1c8845076a","Type":"ContainerStarted","Data":"cd4ed6b8c9437b565d277bc494c92e6a2ccf14fcd15a7666126c833a30b814ee"} Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.018568 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-668f8fbd45-fpqbx" event={"ID":"0fa943f1-8698-41e3-a893-fd1c8845076a","Type":"ContainerStarted","Data":"a08c805cfb747713e43c948ab056929bf738fd7bb29b13b30b8ecd04da5893d0"} Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.024713 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-qs9s6" event={"ID":"ae2f3329-33ca-402d-b841-8f1e257bbf8a","Type":"ContainerStarted","Data":"71b5753eee0c71364cb76a4843975d6ba07a09e3c6338c5ac66b32b4a82d65d0"} Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.024868 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5784cf869f-qs9s6" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.027640 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-67bf9f4b94-nk9h8" event={"ID":"ae9a65ac-f1a7-4ef0-aeba-aca57985ae87","Type":"ContainerStarted","Data":"8427b196183007acae66bef5acceb2a7907bd3146b26f0a899beeb88f811ca69"} Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.027716 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-67bf9f4b94-nk9h8" event={"ID":"ae9a65ac-f1a7-4ef0-aeba-aca57985ae87","Type":"ContainerStarted","Data":"95a000e928ecdaabc42672ae17d033350ed801320965e4a4040c3067b860c6fd"} Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.031840 4840 generic.go:334] "Generic (PLEG): container finished" podID="e4a730c4-623d-4ced-a687-bec2d627eae9" containerID="19c1bb2cbbc34614446d73b508c35aafb815b7a569fbfdb49916e9cdaadb3519" exitCode=0 Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.031916 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e4a730c4-623d-4ced-a687-bec2d627eae9","Type":"ContainerDied","Data":"19c1bb2cbbc34614446d73b508c35aafb815b7a569fbfdb49916e9cdaadb3519"} Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.034760 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3","Type":"ContainerStarted","Data":"4c101ba43c368a4ff85926c05dc7c575d9680e0252a1474b752fee17d9cb5e1d"} Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.040733 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"22734746-33bf-495a-9c55-b7c4cf274fa7","Type":"ContainerStarted","Data":"5fab445fa149d9b2553cdc4abfb2b0ebeb620d47d6f46a1c098b269b17463467"} Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.041655 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.047682 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-668f8fbd45-fpqbx" podStartSLOduration=4.497472182 podStartE2EDuration="8.047658094s" podCreationTimestamp="2025-09-30 14:13:27 +0000 UTC" firstStartedPulling="2025-09-30 14:13:30.271746218 +0000 UTC m=+1038.900832641" lastFinishedPulling="2025-09-30 14:13:33.82193213 +0000 UTC m=+1042.451018553" observedRunningTime="2025-09-30 14:13:35.033662909 +0000 UTC m=+1043.662749352" watchObservedRunningTime="2025-09-30 14:13:35.047658094 +0000 UTC m=+1043.676744527" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.062563 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-67bf9f4b94-nk9h8" podStartSLOduration=4.61815919 podStartE2EDuration="8.062525552s" podCreationTimestamp="2025-09-30 14:13:27 +0000 UTC" firstStartedPulling="2025-09-30 14:13:30.380605037 +0000 UTC m=+1039.009691460" lastFinishedPulling="2025-09-30 14:13:33.824971399 +0000 UTC m=+1042.454057822" observedRunningTime="2025-09-30 14:13:35.053935208 +0000 UTC m=+1043.683021631" watchObservedRunningTime="2025-09-30 14:13:35.062525552 +0000 UTC m=+1043.691611975" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.092940 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5784cf869f-qs9s6" podStartSLOduration=5.092916624 podStartE2EDuration="5.092916624s" podCreationTimestamp="2025-09-30 14:13:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:13:35.080335046 +0000 UTC m=+1043.709421469" watchObservedRunningTime="2025-09-30 14:13:35.092916624 +0000 UTC m=+1043.722003047" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.118135 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.118119731 podStartE2EDuration="5.118119731s" podCreationTimestamp="2025-09-30 14:13:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:13:35.108159122 +0000 UTC m=+1043.737245545" watchObservedRunningTime="2025-09-30 14:13:35.118119731 +0000 UTC m=+1043.747206154" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.138000 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-784f485bd4-gztrl"] Sep 30 14:13:35 crc kubenswrapper[4840]: E0930 14:13:35.138577 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9947cc2-023b-407f-a6c2-b4b2e2c67455" containerName="init" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.138589 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9947cc2-023b-407f-a6c2-b4b2e2c67455" containerName="init" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.138785 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9947cc2-023b-407f-a6c2-b4b2e2c67455" containerName="init" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.139706 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-784f485bd4-gztrl" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.143725 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-784f485bd4-gztrl"] Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.144692 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.144877 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.234248 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqwp5\" (UniqueName: \"kubernetes.io/projected/f7f068af-286b-4d75-ace9-247f0432de8a-kube-api-access-dqwp5\") pod \"barbican-api-784f485bd4-gztrl\" (UID: \"f7f068af-286b-4d75-ace9-247f0432de8a\") " pod="openstack/barbican-api-784f485bd4-gztrl" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.234292 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7f068af-286b-4d75-ace9-247f0432de8a-public-tls-certs\") pod \"barbican-api-784f485bd4-gztrl\" (UID: \"f7f068af-286b-4d75-ace9-247f0432de8a\") " pod="openstack/barbican-api-784f485bd4-gztrl" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.234318 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f7f068af-286b-4d75-ace9-247f0432de8a-config-data-custom\") pod \"barbican-api-784f485bd4-gztrl\" (UID: \"f7f068af-286b-4d75-ace9-247f0432de8a\") " pod="openstack/barbican-api-784f485bd4-gztrl" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.234520 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7f068af-286b-4d75-ace9-247f0432de8a-internal-tls-certs\") pod \"barbican-api-784f485bd4-gztrl\" (UID: \"f7f068af-286b-4d75-ace9-247f0432de8a\") " pod="openstack/barbican-api-784f485bd4-gztrl" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.234575 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7f068af-286b-4d75-ace9-247f0432de8a-config-data\") pod \"barbican-api-784f485bd4-gztrl\" (UID: \"f7f068af-286b-4d75-ace9-247f0432de8a\") " pod="openstack/barbican-api-784f485bd4-gztrl" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.234745 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7f068af-286b-4d75-ace9-247f0432de8a-combined-ca-bundle\") pod \"barbican-api-784f485bd4-gztrl\" (UID: \"f7f068af-286b-4d75-ace9-247f0432de8a\") " pod="openstack/barbican-api-784f485bd4-gztrl" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.234767 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7f068af-286b-4d75-ace9-247f0432de8a-logs\") pod \"barbican-api-784f485bd4-gztrl\" (UID: \"f7f068af-286b-4d75-ace9-247f0432de8a\") " pod="openstack/barbican-api-784f485bd4-gztrl" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.336695 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqwp5\" (UniqueName: \"kubernetes.io/projected/f7f068af-286b-4d75-ace9-247f0432de8a-kube-api-access-dqwp5\") pod \"barbican-api-784f485bd4-gztrl\" (UID: \"f7f068af-286b-4d75-ace9-247f0432de8a\") " pod="openstack/barbican-api-784f485bd4-gztrl" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.336747 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7f068af-286b-4d75-ace9-247f0432de8a-public-tls-certs\") pod \"barbican-api-784f485bd4-gztrl\" (UID: \"f7f068af-286b-4d75-ace9-247f0432de8a\") " pod="openstack/barbican-api-784f485bd4-gztrl" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.336770 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f7f068af-286b-4d75-ace9-247f0432de8a-config-data-custom\") pod \"barbican-api-784f485bd4-gztrl\" (UID: \"f7f068af-286b-4d75-ace9-247f0432de8a\") " pod="openstack/barbican-api-784f485bd4-gztrl" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.336814 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7f068af-286b-4d75-ace9-247f0432de8a-internal-tls-certs\") pod \"barbican-api-784f485bd4-gztrl\" (UID: \"f7f068af-286b-4d75-ace9-247f0432de8a\") " pod="openstack/barbican-api-784f485bd4-gztrl" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.336832 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7f068af-286b-4d75-ace9-247f0432de8a-config-data\") pod \"barbican-api-784f485bd4-gztrl\" (UID: \"f7f068af-286b-4d75-ace9-247f0432de8a\") " pod="openstack/barbican-api-784f485bd4-gztrl" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.336892 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7f068af-286b-4d75-ace9-247f0432de8a-combined-ca-bundle\") pod \"barbican-api-784f485bd4-gztrl\" (UID: \"f7f068af-286b-4d75-ace9-247f0432de8a\") " pod="openstack/barbican-api-784f485bd4-gztrl" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.336908 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7f068af-286b-4d75-ace9-247f0432de8a-logs\") pod \"barbican-api-784f485bd4-gztrl\" (UID: \"f7f068af-286b-4d75-ace9-247f0432de8a\") " pod="openstack/barbican-api-784f485bd4-gztrl" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.337381 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7f068af-286b-4d75-ace9-247f0432de8a-logs\") pod \"barbican-api-784f485bd4-gztrl\" (UID: \"f7f068af-286b-4d75-ace9-247f0432de8a\") " pod="openstack/barbican-api-784f485bd4-gztrl" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.346903 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7f068af-286b-4d75-ace9-247f0432de8a-config-data\") pod \"barbican-api-784f485bd4-gztrl\" (UID: \"f7f068af-286b-4d75-ace9-247f0432de8a\") " pod="openstack/barbican-api-784f485bd4-gztrl" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.347470 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7f068af-286b-4d75-ace9-247f0432de8a-public-tls-certs\") pod \"barbican-api-784f485bd4-gztrl\" (UID: \"f7f068af-286b-4d75-ace9-247f0432de8a\") " pod="openstack/barbican-api-784f485bd4-gztrl" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.347631 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7f068af-286b-4d75-ace9-247f0432de8a-internal-tls-certs\") pod \"barbican-api-784f485bd4-gztrl\" (UID: \"f7f068af-286b-4d75-ace9-247f0432de8a\") " pod="openstack/barbican-api-784f485bd4-gztrl" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.352219 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f7f068af-286b-4d75-ace9-247f0432de8a-config-data-custom\") pod \"barbican-api-784f485bd4-gztrl\" (UID: \"f7f068af-286b-4d75-ace9-247f0432de8a\") " pod="openstack/barbican-api-784f485bd4-gztrl" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.355606 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqwp5\" (UniqueName: \"kubernetes.io/projected/f7f068af-286b-4d75-ace9-247f0432de8a-kube-api-access-dqwp5\") pod \"barbican-api-784f485bd4-gztrl\" (UID: \"f7f068af-286b-4d75-ace9-247f0432de8a\") " pod="openstack/barbican-api-784f485bd4-gztrl" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.368914 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7f068af-286b-4d75-ace9-247f0432de8a-combined-ca-bundle\") pod \"barbican-api-784f485bd4-gztrl\" (UID: \"f7f068af-286b-4d75-ace9-247f0432de8a\") " pod="openstack/barbican-api-784f485bd4-gztrl" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.463158 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.487242 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-784f485bd4-gztrl" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.539227 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4a730c4-623d-4ced-a687-bec2d627eae9-scripts\") pod \"e4a730c4-623d-4ced-a687-bec2d627eae9\" (UID: \"e4a730c4-623d-4ced-a687-bec2d627eae9\") " Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.539317 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e4a730c4-623d-4ced-a687-bec2d627eae9-sg-core-conf-yaml\") pod \"e4a730c4-623d-4ced-a687-bec2d627eae9\" (UID: \"e4a730c4-623d-4ced-a687-bec2d627eae9\") " Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.539398 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e4a730c4-623d-4ced-a687-bec2d627eae9-run-httpd\") pod \"e4a730c4-623d-4ced-a687-bec2d627eae9\" (UID: \"e4a730c4-623d-4ced-a687-bec2d627eae9\") " Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.539445 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e4a730c4-623d-4ced-a687-bec2d627eae9-log-httpd\") pod \"e4a730c4-623d-4ced-a687-bec2d627eae9\" (UID: \"e4a730c4-623d-4ced-a687-bec2d627eae9\") " Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.539483 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hnb5k\" (UniqueName: \"kubernetes.io/projected/e4a730c4-623d-4ced-a687-bec2d627eae9-kube-api-access-hnb5k\") pod \"e4a730c4-623d-4ced-a687-bec2d627eae9\" (UID: \"e4a730c4-623d-4ced-a687-bec2d627eae9\") " Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.539518 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4a730c4-623d-4ced-a687-bec2d627eae9-combined-ca-bundle\") pod \"e4a730c4-623d-4ced-a687-bec2d627eae9\" (UID: \"e4a730c4-623d-4ced-a687-bec2d627eae9\") " Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.539607 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4a730c4-623d-4ced-a687-bec2d627eae9-config-data\") pod \"e4a730c4-623d-4ced-a687-bec2d627eae9\" (UID: \"e4a730c4-623d-4ced-a687-bec2d627eae9\") " Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.540888 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4a730c4-623d-4ced-a687-bec2d627eae9-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e4a730c4-623d-4ced-a687-bec2d627eae9" (UID: "e4a730c4-623d-4ced-a687-bec2d627eae9"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.540923 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4a730c4-623d-4ced-a687-bec2d627eae9-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e4a730c4-623d-4ced-a687-bec2d627eae9" (UID: "e4a730c4-623d-4ced-a687-bec2d627eae9"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.542973 4840 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e4a730c4-623d-4ced-a687-bec2d627eae9-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.543000 4840 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e4a730c4-623d-4ced-a687-bec2d627eae9-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.543011 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4a730c4-623d-4ced-a687-bec2d627eae9-scripts" (OuterVolumeSpecName: "scripts") pod "e4a730c4-623d-4ced-a687-bec2d627eae9" (UID: "e4a730c4-623d-4ced-a687-bec2d627eae9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.551961 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4a730c4-623d-4ced-a687-bec2d627eae9-kube-api-access-hnb5k" (OuterVolumeSpecName: "kube-api-access-hnb5k") pod "e4a730c4-623d-4ced-a687-bec2d627eae9" (UID: "e4a730c4-623d-4ced-a687-bec2d627eae9"). InnerVolumeSpecName "kube-api-access-hnb5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.567997 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4a730c4-623d-4ced-a687-bec2d627eae9-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e4a730c4-623d-4ced-a687-bec2d627eae9" (UID: "e4a730c4-623d-4ced-a687-bec2d627eae9"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.645270 4840 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4a730c4-623d-4ced-a687-bec2d627eae9-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.645295 4840 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e4a730c4-623d-4ced-a687-bec2d627eae9-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.645305 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hnb5k\" (UniqueName: \"kubernetes.io/projected/e4a730c4-623d-4ced-a687-bec2d627eae9-kube-api-access-hnb5k\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.652775 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4a730c4-623d-4ced-a687-bec2d627eae9-config-data" (OuterVolumeSpecName: "config-data") pod "e4a730c4-623d-4ced-a687-bec2d627eae9" (UID: "e4a730c4-623d-4ced-a687-bec2d627eae9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.657333 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4a730c4-623d-4ced-a687-bec2d627eae9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e4a730c4-623d-4ced-a687-bec2d627eae9" (UID: "e4a730c4-623d-4ced-a687-bec2d627eae9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.746651 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4a730c4-623d-4ced-a687-bec2d627eae9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:35 crc kubenswrapper[4840]: I0930 14:13:35.746699 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4a730c4-623d-4ced-a687-bec2d627eae9-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:35.999855 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-784f485bd4-gztrl"] Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.060438 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e4a730c4-623d-4ced-a687-bec2d627eae9","Type":"ContainerDied","Data":"d36952b559a81c75053f08f503612e3c98121ccd14059d960e29d7496be69ca8"} Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.060485 4840 scope.go:117] "RemoveContainer" containerID="8f4c2238d84a252e99a591f48de599db61ae31531f66f05d1cefccc93adf1543" Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.060613 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.139928 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="22734746-33bf-495a-9c55-b7c4cf274fa7" containerName="cinder-api-log" containerID="cri-o://80927d30a386a824baebf38a057bc97ccb64ade654bbc94b57438f9bbafee003" gracePeriod=30 Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.140123 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="22734746-33bf-495a-9c55-b7c4cf274fa7" containerName="cinder-api" containerID="cri-o://5fab445fa149d9b2553cdc4abfb2b0ebeb620d47d6f46a1c098b269b17463467" gracePeriod=30 Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.150345 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.150389 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3","Type":"ContainerStarted","Data":"4f6c20bdac5506dafa0cb9b4c126a0ebee27fdd4eea09653d408e0ecf420fbbe"} Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.150415 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-784f485bd4-gztrl" event={"ID":"f7f068af-286b-4d75-ace9-247f0432de8a","Type":"ContainerStarted","Data":"ff28cfece29b33e30045c42f4574142d20fd2cd382a8c7eef7061840ca80a3c7"} Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.154744 4840 scope.go:117] "RemoveContainer" containerID="f2ce0409cad8baf6e2ed54568f34ac02c6b59d18905648cee072364d2def57b6" Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.168628 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.182370 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:13:36 crc kubenswrapper[4840]: E0930 14:13:36.182777 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4a730c4-623d-4ced-a687-bec2d627eae9" containerName="ceilometer-notification-agent" Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.182791 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4a730c4-623d-4ced-a687-bec2d627eae9" containerName="ceilometer-notification-agent" Sep 30 14:13:36 crc kubenswrapper[4840]: E0930 14:13:36.182811 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4a730c4-623d-4ced-a687-bec2d627eae9" containerName="proxy-httpd" Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.182817 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4a730c4-623d-4ced-a687-bec2d627eae9" containerName="proxy-httpd" Sep 30 14:13:36 crc kubenswrapper[4840]: E0930 14:13:36.182854 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4a730c4-623d-4ced-a687-bec2d627eae9" containerName="sg-core" Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.182860 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4a730c4-623d-4ced-a687-bec2d627eae9" containerName="sg-core" Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.183016 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4a730c4-623d-4ced-a687-bec2d627eae9" containerName="ceilometer-notification-agent" Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.183035 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4a730c4-623d-4ced-a687-bec2d627eae9" containerName="sg-core" Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.183050 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4a730c4-623d-4ced-a687-bec2d627eae9" containerName="proxy-httpd" Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.184742 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.183024 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.23149911 podStartE2EDuration="6.183004852s" podCreationTimestamp="2025-09-30 14:13:30 +0000 UTC" firstStartedPulling="2025-09-30 14:13:31.874416662 +0000 UTC m=+1040.503503085" lastFinishedPulling="2025-09-30 14:13:33.825922404 +0000 UTC m=+1042.455008827" observedRunningTime="2025-09-30 14:13:36.169179621 +0000 UTC m=+1044.798266064" watchObservedRunningTime="2025-09-30 14:13:36.183004852 +0000 UTC m=+1044.812091265" Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.200861 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.201128 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.216512 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.251706 4840 scope.go:117] "RemoveContainer" containerID="19c1bb2cbbc34614446d73b508c35aafb815b7a569fbfdb49916e9cdaadb3519" Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.259100 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9338f796-fcc1-4b71-8c3e-d48d5234d83d-scripts\") pod \"ceilometer-0\" (UID: \"9338f796-fcc1-4b71-8c3e-d48d5234d83d\") " pod="openstack/ceilometer-0" Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.259137 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9338f796-fcc1-4b71-8c3e-d48d5234d83d-config-data\") pod \"ceilometer-0\" (UID: \"9338f796-fcc1-4b71-8c3e-d48d5234d83d\") " pod="openstack/ceilometer-0" Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.259198 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9338f796-fcc1-4b71-8c3e-d48d5234d83d-log-httpd\") pod \"ceilometer-0\" (UID: \"9338f796-fcc1-4b71-8c3e-d48d5234d83d\") " pod="openstack/ceilometer-0" Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.259290 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9338f796-fcc1-4b71-8c3e-d48d5234d83d-run-httpd\") pod \"ceilometer-0\" (UID: \"9338f796-fcc1-4b71-8c3e-d48d5234d83d\") " pod="openstack/ceilometer-0" Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.259310 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9338f796-fcc1-4b71-8c3e-d48d5234d83d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9338f796-fcc1-4b71-8c3e-d48d5234d83d\") " pod="openstack/ceilometer-0" Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.259328 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdqzz\" (UniqueName: \"kubernetes.io/projected/9338f796-fcc1-4b71-8c3e-d48d5234d83d-kube-api-access-fdqzz\") pod \"ceilometer-0\" (UID: \"9338f796-fcc1-4b71-8c3e-d48d5234d83d\") " pod="openstack/ceilometer-0" Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.259344 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9338f796-fcc1-4b71-8c3e-d48d5234d83d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9338f796-fcc1-4b71-8c3e-d48d5234d83d\") " pod="openstack/ceilometer-0" Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.360483 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9338f796-fcc1-4b71-8c3e-d48d5234d83d-log-httpd\") pod \"ceilometer-0\" (UID: \"9338f796-fcc1-4b71-8c3e-d48d5234d83d\") " pod="openstack/ceilometer-0" Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.360581 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9338f796-fcc1-4b71-8c3e-d48d5234d83d-run-httpd\") pod \"ceilometer-0\" (UID: \"9338f796-fcc1-4b71-8c3e-d48d5234d83d\") " pod="openstack/ceilometer-0" Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.360604 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9338f796-fcc1-4b71-8c3e-d48d5234d83d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9338f796-fcc1-4b71-8c3e-d48d5234d83d\") " pod="openstack/ceilometer-0" Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.360625 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdqzz\" (UniqueName: \"kubernetes.io/projected/9338f796-fcc1-4b71-8c3e-d48d5234d83d-kube-api-access-fdqzz\") pod \"ceilometer-0\" (UID: \"9338f796-fcc1-4b71-8c3e-d48d5234d83d\") " pod="openstack/ceilometer-0" Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.360641 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9338f796-fcc1-4b71-8c3e-d48d5234d83d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9338f796-fcc1-4b71-8c3e-d48d5234d83d\") " pod="openstack/ceilometer-0" Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.360737 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9338f796-fcc1-4b71-8c3e-d48d5234d83d-scripts\") pod \"ceilometer-0\" (UID: \"9338f796-fcc1-4b71-8c3e-d48d5234d83d\") " pod="openstack/ceilometer-0" Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.360751 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9338f796-fcc1-4b71-8c3e-d48d5234d83d-config-data\") pod \"ceilometer-0\" (UID: \"9338f796-fcc1-4b71-8c3e-d48d5234d83d\") " pod="openstack/ceilometer-0" Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.361641 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9338f796-fcc1-4b71-8c3e-d48d5234d83d-log-httpd\") pod \"ceilometer-0\" (UID: \"9338f796-fcc1-4b71-8c3e-d48d5234d83d\") " pod="openstack/ceilometer-0" Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.361962 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9338f796-fcc1-4b71-8c3e-d48d5234d83d-run-httpd\") pod \"ceilometer-0\" (UID: \"9338f796-fcc1-4b71-8c3e-d48d5234d83d\") " pod="openstack/ceilometer-0" Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.364905 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9338f796-fcc1-4b71-8c3e-d48d5234d83d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9338f796-fcc1-4b71-8c3e-d48d5234d83d\") " pod="openstack/ceilometer-0" Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.365342 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9338f796-fcc1-4b71-8c3e-d48d5234d83d-scripts\") pod \"ceilometer-0\" (UID: \"9338f796-fcc1-4b71-8c3e-d48d5234d83d\") " pod="openstack/ceilometer-0" Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.367141 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9338f796-fcc1-4b71-8c3e-d48d5234d83d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9338f796-fcc1-4b71-8c3e-d48d5234d83d\") " pod="openstack/ceilometer-0" Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.371726 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9338f796-fcc1-4b71-8c3e-d48d5234d83d-config-data\") pod \"ceilometer-0\" (UID: \"9338f796-fcc1-4b71-8c3e-d48d5234d83d\") " pod="openstack/ceilometer-0" Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.377745 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdqzz\" (UniqueName: \"kubernetes.io/projected/9338f796-fcc1-4b71-8c3e-d48d5234d83d-kube-api-access-fdqzz\") pod \"ceilometer-0\" (UID: \"9338f796-fcc1-4b71-8c3e-d48d5234d83d\") " pod="openstack/ceilometer-0" Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.553151 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:13:36 crc kubenswrapper[4840]: I0930 14:13:36.992877 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.014235 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.080999 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzxfv\" (UniqueName: \"kubernetes.io/projected/22734746-33bf-495a-9c55-b7c4cf274fa7-kube-api-access-xzxfv\") pod \"22734746-33bf-495a-9c55-b7c4cf274fa7\" (UID: \"22734746-33bf-495a-9c55-b7c4cf274fa7\") " Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.081090 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/22734746-33bf-495a-9c55-b7c4cf274fa7-config-data-custom\") pod \"22734746-33bf-495a-9c55-b7c4cf274fa7\" (UID: \"22734746-33bf-495a-9c55-b7c4cf274fa7\") " Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.081623 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22734746-33bf-495a-9c55-b7c4cf274fa7-combined-ca-bundle\") pod \"22734746-33bf-495a-9c55-b7c4cf274fa7\" (UID: \"22734746-33bf-495a-9c55-b7c4cf274fa7\") " Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.081670 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22734746-33bf-495a-9c55-b7c4cf274fa7-logs\") pod \"22734746-33bf-495a-9c55-b7c4cf274fa7\" (UID: \"22734746-33bf-495a-9c55-b7c4cf274fa7\") " Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.081766 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/22734746-33bf-495a-9c55-b7c4cf274fa7-etc-machine-id\") pod \"22734746-33bf-495a-9c55-b7c4cf274fa7\" (UID: \"22734746-33bf-495a-9c55-b7c4cf274fa7\") " Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.081789 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22734746-33bf-495a-9c55-b7c4cf274fa7-config-data\") pod \"22734746-33bf-495a-9c55-b7c4cf274fa7\" (UID: \"22734746-33bf-495a-9c55-b7c4cf274fa7\") " Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.081811 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22734746-33bf-495a-9c55-b7c4cf274fa7-scripts\") pod \"22734746-33bf-495a-9c55-b7c4cf274fa7\" (UID: \"22734746-33bf-495a-9c55-b7c4cf274fa7\") " Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.082391 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/22734746-33bf-495a-9c55-b7c4cf274fa7-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "22734746-33bf-495a-9c55-b7c4cf274fa7" (UID: "22734746-33bf-495a-9c55-b7c4cf274fa7"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.082707 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22734746-33bf-495a-9c55-b7c4cf274fa7-logs" (OuterVolumeSpecName: "logs") pod "22734746-33bf-495a-9c55-b7c4cf274fa7" (UID: "22734746-33bf-495a-9c55-b7c4cf274fa7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.087775 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22734746-33bf-495a-9c55-b7c4cf274fa7-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "22734746-33bf-495a-9c55-b7c4cf274fa7" (UID: "22734746-33bf-495a-9c55-b7c4cf274fa7"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.087871 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22734746-33bf-495a-9c55-b7c4cf274fa7-scripts" (OuterVolumeSpecName: "scripts") pod "22734746-33bf-495a-9c55-b7c4cf274fa7" (UID: "22734746-33bf-495a-9c55-b7c4cf274fa7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.088045 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22734746-33bf-495a-9c55-b7c4cf274fa7-kube-api-access-xzxfv" (OuterVolumeSpecName: "kube-api-access-xzxfv") pod "22734746-33bf-495a-9c55-b7c4cf274fa7" (UID: "22734746-33bf-495a-9c55-b7c4cf274fa7"). InnerVolumeSpecName "kube-api-access-xzxfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.113221 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22734746-33bf-495a-9c55-b7c4cf274fa7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "22734746-33bf-495a-9c55-b7c4cf274fa7" (UID: "22734746-33bf-495a-9c55-b7c4cf274fa7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.131564 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22734746-33bf-495a-9c55-b7c4cf274fa7-config-data" (OuterVolumeSpecName: "config-data") pod "22734746-33bf-495a-9c55-b7c4cf274fa7" (UID: "22734746-33bf-495a-9c55-b7c4cf274fa7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.152146 4840 generic.go:334] "Generic (PLEG): container finished" podID="22734746-33bf-495a-9c55-b7c4cf274fa7" containerID="5fab445fa149d9b2553cdc4abfb2b0ebeb620d47d6f46a1c098b269b17463467" exitCode=0 Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.152174 4840 generic.go:334] "Generic (PLEG): container finished" podID="22734746-33bf-495a-9c55-b7c4cf274fa7" containerID="80927d30a386a824baebf38a057bc97ccb64ade654bbc94b57438f9bbafee003" exitCode=143 Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.152181 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"22734746-33bf-495a-9c55-b7c4cf274fa7","Type":"ContainerDied","Data":"5fab445fa149d9b2553cdc4abfb2b0ebeb620d47d6f46a1c098b269b17463467"} Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.152215 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"22734746-33bf-495a-9c55-b7c4cf274fa7","Type":"ContainerDied","Data":"80927d30a386a824baebf38a057bc97ccb64ade654bbc94b57438f9bbafee003"} Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.152226 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"22734746-33bf-495a-9c55-b7c4cf274fa7","Type":"ContainerDied","Data":"fb02f64294175d668c0e3e208d67842b7558e1f80e91f393ed5924c5a26ee556"} Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.152241 4840 scope.go:117] "RemoveContainer" containerID="5fab445fa149d9b2553cdc4abfb2b0ebeb620d47d6f46a1c098b269b17463467" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.152246 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.155358 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-784f485bd4-gztrl" event={"ID":"f7f068af-286b-4d75-ace9-247f0432de8a","Type":"ContainerStarted","Data":"3314e49ca52aa06828263cfec820b5b1f81221895b180d595ff951bf5dfb565f"} Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.155396 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-784f485bd4-gztrl" event={"ID":"f7f068af-286b-4d75-ace9-247f0432de8a","Type":"ContainerStarted","Data":"8784dc09f1f2530dc46c6c361c3849120a332e46ba3c46ed97b73f58e8e28f3e"} Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.155471 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-784f485bd4-gztrl" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.158755 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9338f796-fcc1-4b71-8c3e-d48d5234d83d","Type":"ContainerStarted","Data":"498887b9fc0395006aaf1d420b15b949d3c567b64f943cc13c1ffc1bc991807f"} Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.179620 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-784f485bd4-gztrl" podStartSLOduration=2.179596381 podStartE2EDuration="2.179596381s" podCreationTimestamp="2025-09-30 14:13:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:13:37.175602677 +0000 UTC m=+1045.804689100" watchObservedRunningTime="2025-09-30 14:13:37.179596381 +0000 UTC m=+1045.808682814" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.179660 4840 scope.go:117] "RemoveContainer" containerID="80927d30a386a824baebf38a057bc97ccb64ade654bbc94b57438f9bbafee003" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.189866 4840 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/22734746-33bf-495a-9c55-b7c4cf274fa7-etc-machine-id\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.189896 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22734746-33bf-495a-9c55-b7c4cf274fa7-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.189904 4840 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22734746-33bf-495a-9c55-b7c4cf274fa7-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.189914 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzxfv\" (UniqueName: \"kubernetes.io/projected/22734746-33bf-495a-9c55-b7c4cf274fa7-kube-api-access-xzxfv\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.189923 4840 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/22734746-33bf-495a-9c55-b7c4cf274fa7-config-data-custom\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.189931 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22734746-33bf-495a-9c55-b7c4cf274fa7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.189939 4840 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22734746-33bf-495a-9c55-b7c4cf274fa7-logs\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.204673 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.204396 4840 scope.go:117] "RemoveContainer" containerID="5fab445fa149d9b2553cdc4abfb2b0ebeb620d47d6f46a1c098b269b17463467" Sep 30 14:13:37 crc kubenswrapper[4840]: E0930 14:13:37.207574 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5fab445fa149d9b2553cdc4abfb2b0ebeb620d47d6f46a1c098b269b17463467\": container with ID starting with 5fab445fa149d9b2553cdc4abfb2b0ebeb620d47d6f46a1c098b269b17463467 not found: ID does not exist" containerID="5fab445fa149d9b2553cdc4abfb2b0ebeb620d47d6f46a1c098b269b17463467" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.207633 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5fab445fa149d9b2553cdc4abfb2b0ebeb620d47d6f46a1c098b269b17463467"} err="failed to get container status \"5fab445fa149d9b2553cdc4abfb2b0ebeb620d47d6f46a1c098b269b17463467\": rpc error: code = NotFound desc = could not find container \"5fab445fa149d9b2553cdc4abfb2b0ebeb620d47d6f46a1c098b269b17463467\": container with ID starting with 5fab445fa149d9b2553cdc4abfb2b0ebeb620d47d6f46a1c098b269b17463467 not found: ID does not exist" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.207684 4840 scope.go:117] "RemoveContainer" containerID="80927d30a386a824baebf38a057bc97ccb64ade654bbc94b57438f9bbafee003" Sep 30 14:13:37 crc kubenswrapper[4840]: E0930 14:13:37.212746 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80927d30a386a824baebf38a057bc97ccb64ade654bbc94b57438f9bbafee003\": container with ID starting with 80927d30a386a824baebf38a057bc97ccb64ade654bbc94b57438f9bbafee003 not found: ID does not exist" containerID="80927d30a386a824baebf38a057bc97ccb64ade654bbc94b57438f9bbafee003" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.212815 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80927d30a386a824baebf38a057bc97ccb64ade654bbc94b57438f9bbafee003"} err="failed to get container status \"80927d30a386a824baebf38a057bc97ccb64ade654bbc94b57438f9bbafee003\": rpc error: code = NotFound desc = could not find container \"80927d30a386a824baebf38a057bc97ccb64ade654bbc94b57438f9bbafee003\": container with ID starting with 80927d30a386a824baebf38a057bc97ccb64ade654bbc94b57438f9bbafee003 not found: ID does not exist" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.212860 4840 scope.go:117] "RemoveContainer" containerID="5fab445fa149d9b2553cdc4abfb2b0ebeb620d47d6f46a1c098b269b17463467" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.215904 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5fab445fa149d9b2553cdc4abfb2b0ebeb620d47d6f46a1c098b269b17463467"} err="failed to get container status \"5fab445fa149d9b2553cdc4abfb2b0ebeb620d47d6f46a1c098b269b17463467\": rpc error: code = NotFound desc = could not find container \"5fab445fa149d9b2553cdc4abfb2b0ebeb620d47d6f46a1c098b269b17463467\": container with ID starting with 5fab445fa149d9b2553cdc4abfb2b0ebeb620d47d6f46a1c098b269b17463467 not found: ID does not exist" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.215968 4840 scope.go:117] "RemoveContainer" containerID="80927d30a386a824baebf38a057bc97ccb64ade654bbc94b57438f9bbafee003" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.216937 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80927d30a386a824baebf38a057bc97ccb64ade654bbc94b57438f9bbafee003"} err="failed to get container status \"80927d30a386a824baebf38a057bc97ccb64ade654bbc94b57438f9bbafee003\": rpc error: code = NotFound desc = could not find container \"80927d30a386a824baebf38a057bc97ccb64ade654bbc94b57438f9bbafee003\": container with ID starting with 80927d30a386a824baebf38a057bc97ccb64ade654bbc94b57438f9bbafee003 not found: ID does not exist" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.218777 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.238293 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Sep 30 14:13:37 crc kubenswrapper[4840]: E0930 14:13:37.238791 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22734746-33bf-495a-9c55-b7c4cf274fa7" containerName="cinder-api-log" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.238811 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="22734746-33bf-495a-9c55-b7c4cf274fa7" containerName="cinder-api-log" Sep 30 14:13:37 crc kubenswrapper[4840]: E0930 14:13:37.238837 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22734746-33bf-495a-9c55-b7c4cf274fa7" containerName="cinder-api" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.238845 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="22734746-33bf-495a-9c55-b7c4cf274fa7" containerName="cinder-api" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.239082 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="22734746-33bf-495a-9c55-b7c4cf274fa7" containerName="cinder-api-log" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.239111 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="22734746-33bf-495a-9c55-b7c4cf274fa7" containerName="cinder-api" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.240309 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.242511 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.242891 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.243226 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.245930 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.292046 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eb01e3a9-61de-4dc8-954b-221d80a675fb-logs\") pod \"cinder-api-0\" (UID: \"eb01e3a9-61de-4dc8-954b-221d80a675fb\") " pod="openstack/cinder-api-0" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.292113 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eb01e3a9-61de-4dc8-954b-221d80a675fb-config-data-custom\") pod \"cinder-api-0\" (UID: \"eb01e3a9-61de-4dc8-954b-221d80a675fb\") " pod="openstack/cinder-api-0" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.292215 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb01e3a9-61de-4dc8-954b-221d80a675fb-config-data\") pod \"cinder-api-0\" (UID: \"eb01e3a9-61de-4dc8-954b-221d80a675fb\") " pod="openstack/cinder-api-0" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.293957 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kndr\" (UniqueName: \"kubernetes.io/projected/eb01e3a9-61de-4dc8-954b-221d80a675fb-kube-api-access-8kndr\") pod \"cinder-api-0\" (UID: \"eb01e3a9-61de-4dc8-954b-221d80a675fb\") " pod="openstack/cinder-api-0" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.294033 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb01e3a9-61de-4dc8-954b-221d80a675fb-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"eb01e3a9-61de-4dc8-954b-221d80a675fb\") " pod="openstack/cinder-api-0" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.294079 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb01e3a9-61de-4dc8-954b-221d80a675fb-scripts\") pod \"cinder-api-0\" (UID: \"eb01e3a9-61de-4dc8-954b-221d80a675fb\") " pod="openstack/cinder-api-0" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.294097 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb01e3a9-61de-4dc8-954b-221d80a675fb-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"eb01e3a9-61de-4dc8-954b-221d80a675fb\") " pod="openstack/cinder-api-0" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.294203 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb01e3a9-61de-4dc8-954b-221d80a675fb-public-tls-certs\") pod \"cinder-api-0\" (UID: \"eb01e3a9-61de-4dc8-954b-221d80a675fb\") " pod="openstack/cinder-api-0" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.294294 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eb01e3a9-61de-4dc8-954b-221d80a675fb-etc-machine-id\") pod \"cinder-api-0\" (UID: \"eb01e3a9-61de-4dc8-954b-221d80a675fb\") " pod="openstack/cinder-api-0" Sep 30 14:13:37 crc kubenswrapper[4840]: W0930 14:13:37.311915 4840 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c465939_fcbe_44b2_b7ae_a24b5a581dbc.slice/crio-conmon-7b37e879a3a49adb8414c595ca1aea847737f1f8f450957b4c08d0a2f64813e5.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c465939_fcbe_44b2_b7ae_a24b5a581dbc.slice/crio-conmon-7b37e879a3a49adb8414c595ca1aea847737f1f8f450957b4c08d0a2f64813e5.scope: no such file or directory Sep 30 14:13:37 crc kubenswrapper[4840]: W0930 14:13:37.311960 4840 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c465939_fcbe_44b2_b7ae_a24b5a581dbc.slice/crio-7b37e879a3a49adb8414c595ca1aea847737f1f8f450957b4c08d0a2f64813e5.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c465939_fcbe_44b2_b7ae_a24b5a581dbc.slice/crio-7b37e879a3a49adb8414c595ca1aea847737f1f8f450957b4c08d0a2f64813e5.scope: no such file or directory Sep 30 14:13:37 crc kubenswrapper[4840]: W0930 14:13:37.316653 4840 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad748b83_4edd_4545_9386_032185c531c5.slice/crio-conmon-8391b03af598c7e9580a0e3ab6d61f6d82ad2ca29aee7293a25f1fb098fad220.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad748b83_4edd_4545_9386_032185c531c5.slice/crio-conmon-8391b03af598c7e9580a0e3ab6d61f6d82ad2ca29aee7293a25f1fb098fad220.scope: no such file or directory Sep 30 14:13:37 crc kubenswrapper[4840]: W0930 14:13:37.316741 4840 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad748b83_4edd_4545_9386_032185c531c5.slice/crio-8391b03af598c7e9580a0e3ab6d61f6d82ad2ca29aee7293a25f1fb098fad220.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad748b83_4edd_4545_9386_032185c531c5.slice/crio-8391b03af598c7e9580a0e3ab6d61f6d82ad2ca29aee7293a25f1fb098fad220.scope: no such file or directory Sep 30 14:13:37 crc kubenswrapper[4840]: W0930 14:13:37.316766 4840 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod242d13bb_6113_421b_83d1_fcccf91a9c68.slice/crio-conmon-58f6da62f61ae533f5086dfb033f031419373cb17ddf4e1ee41d20d765c50f2c.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod242d13bb_6113_421b_83d1_fcccf91a9c68.slice/crio-conmon-58f6da62f61ae533f5086dfb033f031419373cb17ddf4e1ee41d20d765c50f2c.scope: no such file or directory Sep 30 14:13:37 crc kubenswrapper[4840]: W0930 14:13:37.316787 4840 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod242d13bb_6113_421b_83d1_fcccf91a9c68.slice/crio-58f6da62f61ae533f5086dfb033f031419373cb17ddf4e1ee41d20d765c50f2c.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod242d13bb_6113_421b_83d1_fcccf91a9c68.slice/crio-58f6da62f61ae533f5086dfb033f031419373cb17ddf4e1ee41d20d765c50f2c.scope: no such file or directory Sep 30 14:13:37 crc kubenswrapper[4840]: W0930 14:13:37.319765 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4a730c4_623d_4ced_a687_bec2d627eae9.slice/crio-f2ce0409cad8baf6e2ed54568f34ac02c6b59d18905648cee072364d2def57b6.scope WatchSource:0}: Error finding container f2ce0409cad8baf6e2ed54568f34ac02c6b59d18905648cee072364d2def57b6: Status 404 returned error can't find the container with id f2ce0409cad8baf6e2ed54568f34ac02c6b59d18905648cee072364d2def57b6 Sep 30 14:13:37 crc kubenswrapper[4840]: W0930 14:13:37.350658 4840 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb9947cc2_023b_407f_a6c2_b4b2e2c67455.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb9947cc2_023b_407f_a6c2_b4b2e2c67455.slice: no such file or directory Sep 30 14:13:37 crc kubenswrapper[4840]: W0930 14:13:37.359071 4840 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4a730c4_623d_4ced_a687_bec2d627eae9.slice/crio-conmon-8f4c2238d84a252e99a591f48de599db61ae31531f66f05d1cefccc93adf1543.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4a730c4_623d_4ced_a687_bec2d627eae9.slice/crio-conmon-8f4c2238d84a252e99a591f48de599db61ae31531f66f05d1cefccc93adf1543.scope: no such file or directory Sep 30 14:13:37 crc kubenswrapper[4840]: W0930 14:13:37.359124 4840 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4a730c4_623d_4ced_a687_bec2d627eae9.slice/crio-8f4c2238d84a252e99a591f48de599db61ae31531f66f05d1cefccc93adf1543.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4a730c4_623d_4ced_a687_bec2d627eae9.slice/crio-8f4c2238d84a252e99a591f48de599db61ae31531f66f05d1cefccc93adf1543.scope: no such file or directory Sep 30 14:13:37 crc kubenswrapper[4840]: W0930 14:13:37.389644 4840 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod22734746_33bf_495a_9c55_b7c4cf274fa7.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod22734746_33bf_495a_9c55_b7c4cf274fa7.slice: no such file or directory Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.395468 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kndr\" (UniqueName: \"kubernetes.io/projected/eb01e3a9-61de-4dc8-954b-221d80a675fb-kube-api-access-8kndr\") pod \"cinder-api-0\" (UID: \"eb01e3a9-61de-4dc8-954b-221d80a675fb\") " pod="openstack/cinder-api-0" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.395533 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb01e3a9-61de-4dc8-954b-221d80a675fb-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"eb01e3a9-61de-4dc8-954b-221d80a675fb\") " pod="openstack/cinder-api-0" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.395614 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb01e3a9-61de-4dc8-954b-221d80a675fb-scripts\") pod \"cinder-api-0\" (UID: \"eb01e3a9-61de-4dc8-954b-221d80a675fb\") " pod="openstack/cinder-api-0" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.395679 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb01e3a9-61de-4dc8-954b-221d80a675fb-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"eb01e3a9-61de-4dc8-954b-221d80a675fb\") " pod="openstack/cinder-api-0" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.395784 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb01e3a9-61de-4dc8-954b-221d80a675fb-public-tls-certs\") pod \"cinder-api-0\" (UID: \"eb01e3a9-61de-4dc8-954b-221d80a675fb\") " pod="openstack/cinder-api-0" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.395886 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eb01e3a9-61de-4dc8-954b-221d80a675fb-etc-machine-id\") pod \"cinder-api-0\" (UID: \"eb01e3a9-61de-4dc8-954b-221d80a675fb\") " pod="openstack/cinder-api-0" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.396005 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eb01e3a9-61de-4dc8-954b-221d80a675fb-logs\") pod \"cinder-api-0\" (UID: \"eb01e3a9-61de-4dc8-954b-221d80a675fb\") " pod="openstack/cinder-api-0" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.396035 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eb01e3a9-61de-4dc8-954b-221d80a675fb-config-data-custom\") pod \"cinder-api-0\" (UID: \"eb01e3a9-61de-4dc8-954b-221d80a675fb\") " pod="openstack/cinder-api-0" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.396123 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb01e3a9-61de-4dc8-954b-221d80a675fb-config-data\") pod \"cinder-api-0\" (UID: \"eb01e3a9-61de-4dc8-954b-221d80a675fb\") " pod="openstack/cinder-api-0" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.396909 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eb01e3a9-61de-4dc8-954b-221d80a675fb-etc-machine-id\") pod \"cinder-api-0\" (UID: \"eb01e3a9-61de-4dc8-954b-221d80a675fb\") " pod="openstack/cinder-api-0" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.399542 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eb01e3a9-61de-4dc8-954b-221d80a675fb-logs\") pod \"cinder-api-0\" (UID: \"eb01e3a9-61de-4dc8-954b-221d80a675fb\") " pod="openstack/cinder-api-0" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.428740 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb01e3a9-61de-4dc8-954b-221d80a675fb-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"eb01e3a9-61de-4dc8-954b-221d80a675fb\") " pod="openstack/cinder-api-0" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.428749 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eb01e3a9-61de-4dc8-954b-221d80a675fb-config-data-custom\") pod \"cinder-api-0\" (UID: \"eb01e3a9-61de-4dc8-954b-221d80a675fb\") " pod="openstack/cinder-api-0" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.429196 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb01e3a9-61de-4dc8-954b-221d80a675fb-public-tls-certs\") pod \"cinder-api-0\" (UID: \"eb01e3a9-61de-4dc8-954b-221d80a675fb\") " pod="openstack/cinder-api-0" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.429528 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb01e3a9-61de-4dc8-954b-221d80a675fb-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"eb01e3a9-61de-4dc8-954b-221d80a675fb\") " pod="openstack/cinder-api-0" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.429823 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb01e3a9-61de-4dc8-954b-221d80a675fb-scripts\") pod \"cinder-api-0\" (UID: \"eb01e3a9-61de-4dc8-954b-221d80a675fb\") " pod="openstack/cinder-api-0" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.431348 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb01e3a9-61de-4dc8-954b-221d80a675fb-config-data\") pod \"cinder-api-0\" (UID: \"eb01e3a9-61de-4dc8-954b-221d80a675fb\") " pod="openstack/cinder-api-0" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.439169 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kndr\" (UniqueName: \"kubernetes.io/projected/eb01e3a9-61de-4dc8-954b-221d80a675fb-kube-api-access-8kndr\") pod \"cinder-api-0\" (UID: \"eb01e3a9-61de-4dc8-954b-221d80a675fb\") " pod="openstack/cinder-api-0" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.562532 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 30 14:13:37 crc kubenswrapper[4840]: E0930 14:13:37.605725 4840 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4a730c4_623d_4ced_a687_bec2d627eae9.slice/crio-conmon-f2ce0409cad8baf6e2ed54568f34ac02c6b59d18905648cee072364d2def57b6.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad748b83_4edd_4545_9386_032185c531c5.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod242d13bb_6113_421b_83d1_fcccf91a9c68.slice/crio-e7ae2c59a20eb580f9d27eb067fb6d3b0b51bff398919efd4d34d75f936615d9\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb73e88ee_cb25_466d_9860_0dd54a1c1f4c.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c465939_fcbe_44b2_b7ae_a24b5a581dbc.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad748b83_4edd_4545_9386_032185c531c5.slice/crio-82168598b29a794bbf74c3fee4e2676ee2b702525cf44ada0f7498f39206978d\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod47067ffa_fc01_48b4_abcd_4405053eadca.slice/crio-conmon-6d38e955431f3f93cdb79f42366f8efd916449357858ee1e88db0ae9f7808b40.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb73e88ee_cb25_466d_9860_0dd54a1c1f4c.slice/crio-8f1dc45f77092a5778340c020c0952b71db3f7fad8ff85b132ad5c8cc6e7a5c5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c465939_fcbe_44b2_b7ae_a24b5a581dbc.slice/crio-3958b3f15ef2941ffd7ae2a69bdb6a938691349e2395edb10287002e93dd7857\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda7f0b012_f515_434f_919e_5f0409c8e554.slice/crio-807dbc9df37357f57e1252dc8f83d1de62c3336b8cb4150169bcddd5321ae0e3.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod242d13bb_6113_421b_83d1_fcccf91a9c68.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode141523f_ce9b_4626_b155_b0a82a3f6dc6.slice/crio-conmon-91f9f1d766fa7013e2cf435e7a46c86870d1a985c1f7db12ac2763f63a8b3730.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4a730c4_623d_4ced_a687_bec2d627eae9.slice/crio-19c1bb2cbbc34614446d73b508c35aafb815b7a569fbfdb49916e9cdaadb3519.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4a730c4_623d_4ced_a687_bec2d627eae9.slice/crio-d36952b559a81c75053f08f503612e3c98121ccd14059d960e29d7496be69ca8\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4a730c4_623d_4ced_a687_bec2d627eae9.slice/crio-conmon-19c1bb2cbbc34614446d73b508c35aafb815b7a569fbfdb49916e9cdaadb3519.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb73e88ee_cb25_466d_9860_0dd54a1c1f4c.slice/crio-b2c37df544a67982abf51985e156af05ad950bcc25628ce8bd5c54117c6d35eb\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda7f0b012_f515_434f_919e_5f0409c8e554.slice/crio-conmon-807dbc9df37357f57e1252dc8f83d1de62c3336b8cb4150169bcddd5321ae0e3.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda7f0b012_f515_434f_919e_5f0409c8e554.slice/crio-conmon-2edee90f84eb32c953eeb5dfa1fc2729836da354666c490c7c4bac46e34c4409.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb73e88ee_cb25_466d_9860_0dd54a1c1f4c.slice/crio-conmon-8f1dc45f77092a5778340c020c0952b71db3f7fad8ff85b132ad5c8cc6e7a5c5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4a730c4_623d_4ced_a687_bec2d627eae9.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode141523f_ce9b_4626_b155_b0a82a3f6dc6.slice/crio-91f9f1d766fa7013e2cf435e7a46c86870d1a985c1f7db12ac2763f63a8b3730.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode141523f_ce9b_4626_b155_b0a82a3f6dc6.slice/crio-conmon-71b465222c8ca2c7063b1632b64ef4d2b3dbe73248c1b00977324af0871011fc.scope\": RecentStats: unable to find data in memory cache]" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.858512 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-788d88bd6f-l44pw" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.892319 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f785c99fc-fpf4g" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.912175 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fp459\" (UniqueName: \"kubernetes.io/projected/e141523f-ce9b-4626-b155-b0a82a3f6dc6-kube-api-access-fp459\") pod \"e141523f-ce9b-4626-b155-b0a82a3f6dc6\" (UID: \"e141523f-ce9b-4626-b155-b0a82a3f6dc6\") " Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.912332 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e141523f-ce9b-4626-b155-b0a82a3f6dc6-scripts\") pod \"e141523f-ce9b-4626-b155-b0a82a3f6dc6\" (UID: \"e141523f-ce9b-4626-b155-b0a82a3f6dc6\") " Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.912378 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/47067ffa-fc01-48b4-abcd-4405053eadca-horizon-secret-key\") pod \"47067ffa-fc01-48b4-abcd-4405053eadca\" (UID: \"47067ffa-fc01-48b4-abcd-4405053eadca\") " Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.912412 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e141523f-ce9b-4626-b155-b0a82a3f6dc6-horizon-secret-key\") pod \"e141523f-ce9b-4626-b155-b0a82a3f6dc6\" (UID: \"e141523f-ce9b-4626-b155-b0a82a3f6dc6\") " Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.916604 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47067ffa-fc01-48b4-abcd-4405053eadca-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "47067ffa-fc01-48b4-abcd-4405053eadca" (UID: "47067ffa-fc01-48b4-abcd-4405053eadca"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.912474 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e141523f-ce9b-4626-b155-b0a82a3f6dc6-config-data\") pod \"e141523f-ce9b-4626-b155-b0a82a3f6dc6\" (UID: \"e141523f-ce9b-4626-b155-b0a82a3f6dc6\") " Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.917517 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47067ffa-fc01-48b4-abcd-4405053eadca-logs\") pod \"47067ffa-fc01-48b4-abcd-4405053eadca\" (UID: \"47067ffa-fc01-48b4-abcd-4405053eadca\") " Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.917538 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e141523f-ce9b-4626-b155-b0a82a3f6dc6-logs\") pod \"e141523f-ce9b-4626-b155-b0a82a3f6dc6\" (UID: \"e141523f-ce9b-4626-b155-b0a82a3f6dc6\") " Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.917635 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f9fdp\" (UniqueName: \"kubernetes.io/projected/47067ffa-fc01-48b4-abcd-4405053eadca-kube-api-access-f9fdp\") pod \"47067ffa-fc01-48b4-abcd-4405053eadca\" (UID: \"47067ffa-fc01-48b4-abcd-4405053eadca\") " Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.917666 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/47067ffa-fc01-48b4-abcd-4405053eadca-config-data\") pod \"47067ffa-fc01-48b4-abcd-4405053eadca\" (UID: \"47067ffa-fc01-48b4-abcd-4405053eadca\") " Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.917706 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/47067ffa-fc01-48b4-abcd-4405053eadca-scripts\") pod \"47067ffa-fc01-48b4-abcd-4405053eadca\" (UID: \"47067ffa-fc01-48b4-abcd-4405053eadca\") " Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.918333 4840 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/47067ffa-fc01-48b4-abcd-4405053eadca-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.919378 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e141523f-ce9b-4626-b155-b0a82a3f6dc6-logs" (OuterVolumeSpecName: "logs") pod "e141523f-ce9b-4626-b155-b0a82a3f6dc6" (UID: "e141523f-ce9b-4626-b155-b0a82a3f6dc6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.919733 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e141523f-ce9b-4626-b155-b0a82a3f6dc6-kube-api-access-fp459" (OuterVolumeSpecName: "kube-api-access-fp459") pod "e141523f-ce9b-4626-b155-b0a82a3f6dc6" (UID: "e141523f-ce9b-4626-b155-b0a82a3f6dc6"). InnerVolumeSpecName "kube-api-access-fp459". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.919968 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47067ffa-fc01-48b4-abcd-4405053eadca-logs" (OuterVolumeSpecName: "logs") pod "47067ffa-fc01-48b4-abcd-4405053eadca" (UID: "47067ffa-fc01-48b4-abcd-4405053eadca"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.925281 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5cc547b489-2p58b" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.925894 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47067ffa-fc01-48b4-abcd-4405053eadca-kube-api-access-f9fdp" (OuterVolumeSpecName: "kube-api-access-f9fdp") pod "47067ffa-fc01-48b4-abcd-4405053eadca" (UID: "47067ffa-fc01-48b4-abcd-4405053eadca"). InnerVolumeSpecName "kube-api-access-f9fdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.931087 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e141523f-ce9b-4626-b155-b0a82a3f6dc6-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "e141523f-ce9b-4626-b155-b0a82a3f6dc6" (UID: "e141523f-ce9b-4626-b155-b0a82a3f6dc6"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.991244 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47067ffa-fc01-48b4-abcd-4405053eadca-scripts" (OuterVolumeSpecName: "scripts") pod "47067ffa-fc01-48b4-abcd-4405053eadca" (UID: "47067ffa-fc01-48b4-abcd-4405053eadca"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.991695 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e141523f-ce9b-4626-b155-b0a82a3f6dc6-scripts" (OuterVolumeSpecName: "scripts") pod "e141523f-ce9b-4626-b155-b0a82a3f6dc6" (UID: "e141523f-ce9b-4626-b155-b0a82a3f6dc6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:13:37 crc kubenswrapper[4840]: I0930 14:13:37.993166 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47067ffa-fc01-48b4-abcd-4405053eadca-config-data" (OuterVolumeSpecName: "config-data") pod "47067ffa-fc01-48b4-abcd-4405053eadca" (UID: "47067ffa-fc01-48b4-abcd-4405053eadca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.009202 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e141523f-ce9b-4626-b155-b0a82a3f6dc6-config-data" (OuterVolumeSpecName: "config-data") pod "e141523f-ce9b-4626-b155-b0a82a3f6dc6" (UID: "e141523f-ce9b-4626-b155-b0a82a3f6dc6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.019689 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a7f0b012-f515-434f-919e-5f0409c8e554-config-data\") pod \"a7f0b012-f515-434f-919e-5f0409c8e554\" (UID: \"a7f0b012-f515-434f-919e-5f0409c8e554\") " Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.019731 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a7f0b012-f515-434f-919e-5f0409c8e554-horizon-secret-key\") pod \"a7f0b012-f515-434f-919e-5f0409c8e554\" (UID: \"a7f0b012-f515-434f-919e-5f0409c8e554\") " Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.020487 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hnxc2\" (UniqueName: \"kubernetes.io/projected/a7f0b012-f515-434f-919e-5f0409c8e554-kube-api-access-hnxc2\") pod \"a7f0b012-f515-434f-919e-5f0409c8e554\" (UID: \"a7f0b012-f515-434f-919e-5f0409c8e554\") " Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.020600 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7f0b012-f515-434f-919e-5f0409c8e554-logs\") pod \"a7f0b012-f515-434f-919e-5f0409c8e554\" (UID: \"a7f0b012-f515-434f-919e-5f0409c8e554\") " Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.020780 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a7f0b012-f515-434f-919e-5f0409c8e554-scripts\") pod \"a7f0b012-f515-434f-919e-5f0409c8e554\" (UID: \"a7f0b012-f515-434f-919e-5f0409c8e554\") " Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.021257 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7f0b012-f515-434f-919e-5f0409c8e554-logs" (OuterVolumeSpecName: "logs") pod "a7f0b012-f515-434f-919e-5f0409c8e554" (UID: "a7f0b012-f515-434f-919e-5f0409c8e554"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.021531 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fp459\" (UniqueName: \"kubernetes.io/projected/e141523f-ce9b-4626-b155-b0a82a3f6dc6-kube-api-access-fp459\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.021577 4840 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e141523f-ce9b-4626-b155-b0a82a3f6dc6-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.021586 4840 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e141523f-ce9b-4626-b155-b0a82a3f6dc6-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.021595 4840 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7f0b012-f515-434f-919e-5f0409c8e554-logs\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.021602 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e141523f-ce9b-4626-b155-b0a82a3f6dc6-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.021611 4840 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47067ffa-fc01-48b4-abcd-4405053eadca-logs\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.021623 4840 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e141523f-ce9b-4626-b155-b0a82a3f6dc6-logs\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.021634 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f9fdp\" (UniqueName: \"kubernetes.io/projected/47067ffa-fc01-48b4-abcd-4405053eadca-kube-api-access-f9fdp\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.021645 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/47067ffa-fc01-48b4-abcd-4405053eadca-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.021655 4840 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/47067ffa-fc01-48b4-abcd-4405053eadca-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.023204 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7f0b012-f515-434f-919e-5f0409c8e554-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "a7f0b012-f515-434f-919e-5f0409c8e554" (UID: "a7f0b012-f515-434f-919e-5f0409c8e554"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.023782 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7f0b012-f515-434f-919e-5f0409c8e554-kube-api-access-hnxc2" (OuterVolumeSpecName: "kube-api-access-hnxc2") pod "a7f0b012-f515-434f-919e-5f0409c8e554" (UID: "a7f0b012-f515-434f-919e-5f0409c8e554"). InnerVolumeSpecName "kube-api-access-hnxc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.048083 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7f0b012-f515-434f-919e-5f0409c8e554-scripts" (OuterVolumeSpecName: "scripts") pod "a7f0b012-f515-434f-919e-5f0409c8e554" (UID: "a7f0b012-f515-434f-919e-5f0409c8e554"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.050933 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7f0b012-f515-434f-919e-5f0409c8e554-config-data" (OuterVolumeSpecName: "config-data") pod "a7f0b012-f515-434f-919e-5f0409c8e554" (UID: "a7f0b012-f515-434f-919e-5f0409c8e554"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.085888 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Sep 30 14:13:38 crc kubenswrapper[4840]: W0930 14:13:38.087660 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb01e3a9_61de_4dc8_954b_221d80a675fb.slice/crio-9c45f1e4997bf4149d57a8d4af4df199cf44e2e5c7b993f892734eea54c956e5 WatchSource:0}: Error finding container 9c45f1e4997bf4149d57a8d4af4df199cf44e2e5c7b993f892734eea54c956e5: Status 404 returned error can't find the container with id 9c45f1e4997bf4149d57a8d4af4df199cf44e2e5c7b993f892734eea54c956e5 Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.125797 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hnxc2\" (UniqueName: \"kubernetes.io/projected/a7f0b012-f515-434f-919e-5f0409c8e554-kube-api-access-hnxc2\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.125833 4840 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a7f0b012-f515-434f-919e-5f0409c8e554-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.125842 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a7f0b012-f515-434f-919e-5f0409c8e554-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.125851 4840 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a7f0b012-f515-434f-919e-5f0409c8e554-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.129794 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22734746-33bf-495a-9c55-b7c4cf274fa7" path="/var/lib/kubelet/pods/22734746-33bf-495a-9c55-b7c4cf274fa7/volumes" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.130969 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4a730c4-623d-4ced-a687-bec2d627eae9" path="/var/lib/kubelet/pods/e4a730c4-623d-4ced-a687-bec2d627eae9/volumes" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.171722 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"eb01e3a9-61de-4dc8-954b-221d80a675fb","Type":"ContainerStarted","Data":"9c45f1e4997bf4149d57a8d4af4df199cf44e2e5c7b993f892734eea54c956e5"} Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.174577 4840 generic.go:334] "Generic (PLEG): container finished" podID="a7f0b012-f515-434f-919e-5f0409c8e554" containerID="2edee90f84eb32c953eeb5dfa1fc2729836da354666c490c7c4bac46e34c4409" exitCode=137 Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.174602 4840 generic.go:334] "Generic (PLEG): container finished" podID="a7f0b012-f515-434f-919e-5f0409c8e554" containerID="807dbc9df37357f57e1252dc8f83d1de62c3336b8cb4150169bcddd5321ae0e3" exitCode=137 Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.174634 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5cc547b489-2p58b" event={"ID":"a7f0b012-f515-434f-919e-5f0409c8e554","Type":"ContainerDied","Data":"2edee90f84eb32c953eeb5dfa1fc2729836da354666c490c7c4bac46e34c4409"} Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.174653 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5cc547b489-2p58b" event={"ID":"a7f0b012-f515-434f-919e-5f0409c8e554","Type":"ContainerDied","Data":"807dbc9df37357f57e1252dc8f83d1de62c3336b8cb4150169bcddd5321ae0e3"} Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.174666 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5cc547b489-2p58b" event={"ID":"a7f0b012-f515-434f-919e-5f0409c8e554","Type":"ContainerDied","Data":"a63a10d7ebc945cd9f18a75e8c7429c4f157b2e5fdb85475b333d02a5f7fce57"} Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.174680 4840 scope.go:117] "RemoveContainer" containerID="2edee90f84eb32c953eeb5dfa1fc2729836da354666c490c7c4bac46e34c4409" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.174797 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5cc547b489-2p58b" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.177961 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9338f796-fcc1-4b71-8c3e-d48d5234d83d","Type":"ContainerStarted","Data":"c0b92bbe8cf99b3b221a23959cc07d7ec9173afe14a9f6f6f88255772435471a"} Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.184060 4840 generic.go:334] "Generic (PLEG): container finished" podID="e141523f-ce9b-4626-b155-b0a82a3f6dc6" containerID="71b465222c8ca2c7063b1632b64ef4d2b3dbe73248c1b00977324af0871011fc" exitCode=137 Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.184093 4840 generic.go:334] "Generic (PLEG): container finished" podID="e141523f-ce9b-4626-b155-b0a82a3f6dc6" containerID="91f9f1d766fa7013e2cf435e7a46c86870d1a985c1f7db12ac2763f63a8b3730" exitCode=137 Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.184126 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-788d88bd6f-l44pw" event={"ID":"e141523f-ce9b-4626-b155-b0a82a3f6dc6","Type":"ContainerDied","Data":"71b465222c8ca2c7063b1632b64ef4d2b3dbe73248c1b00977324af0871011fc"} Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.184148 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-788d88bd6f-l44pw" event={"ID":"e141523f-ce9b-4626-b155-b0a82a3f6dc6","Type":"ContainerDied","Data":"91f9f1d766fa7013e2cf435e7a46c86870d1a985c1f7db12ac2763f63a8b3730"} Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.184168 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-788d88bd6f-l44pw" event={"ID":"e141523f-ce9b-4626-b155-b0a82a3f6dc6","Type":"ContainerDied","Data":"99172bb3072324501f5d649ac3715c1f111ca64410df87296a892bccd91a51b5"} Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.184341 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-788d88bd6f-l44pw" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.195787 4840 generic.go:334] "Generic (PLEG): container finished" podID="47067ffa-fc01-48b4-abcd-4405053eadca" containerID="6d38e955431f3f93cdb79f42366f8efd916449357858ee1e88db0ae9f7808b40" exitCode=137 Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.195815 4840 generic.go:334] "Generic (PLEG): container finished" podID="47067ffa-fc01-48b4-abcd-4405053eadca" containerID="1e4111d6b9033739a506b5a02fccfc6924f5a5fd075ba04c56d92cc8bc086c9e" exitCode=137 Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.196591 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f785c99fc-fpf4g" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.196838 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f785c99fc-fpf4g" event={"ID":"47067ffa-fc01-48b4-abcd-4405053eadca","Type":"ContainerDied","Data":"6d38e955431f3f93cdb79f42366f8efd916449357858ee1e88db0ae9f7808b40"} Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.196869 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f785c99fc-fpf4g" event={"ID":"47067ffa-fc01-48b4-abcd-4405053eadca","Type":"ContainerDied","Data":"1e4111d6b9033739a506b5a02fccfc6924f5a5fd075ba04c56d92cc8bc086c9e"} Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.196879 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f785c99fc-fpf4g" event={"ID":"47067ffa-fc01-48b4-abcd-4405053eadca","Type":"ContainerDied","Data":"6102f664417c2451768375070aac7093a8c91d3a91dbc55122b2021866de56ca"} Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.196933 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-784f485bd4-gztrl" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.198946 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5cc547b489-2p58b"] Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.208685 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5cc547b489-2p58b"] Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.230137 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-788d88bd6f-l44pw"] Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.238564 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-788d88bd6f-l44pw"] Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.247540 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7f785c99fc-fpf4g"] Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.256729 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7f785c99fc-fpf4g"] Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.386376 4840 scope.go:117] "RemoveContainer" containerID="807dbc9df37357f57e1252dc8f83d1de62c3336b8cb4150169bcddd5321ae0e3" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.591591 4840 scope.go:117] "RemoveContainer" containerID="2edee90f84eb32c953eeb5dfa1fc2729836da354666c490c7c4bac46e34c4409" Sep 30 14:13:38 crc kubenswrapper[4840]: E0930 14:13:38.592354 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2edee90f84eb32c953eeb5dfa1fc2729836da354666c490c7c4bac46e34c4409\": container with ID starting with 2edee90f84eb32c953eeb5dfa1fc2729836da354666c490c7c4bac46e34c4409 not found: ID does not exist" containerID="2edee90f84eb32c953eeb5dfa1fc2729836da354666c490c7c4bac46e34c4409" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.592393 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2edee90f84eb32c953eeb5dfa1fc2729836da354666c490c7c4bac46e34c4409"} err="failed to get container status \"2edee90f84eb32c953eeb5dfa1fc2729836da354666c490c7c4bac46e34c4409\": rpc error: code = NotFound desc = could not find container \"2edee90f84eb32c953eeb5dfa1fc2729836da354666c490c7c4bac46e34c4409\": container with ID starting with 2edee90f84eb32c953eeb5dfa1fc2729836da354666c490c7c4bac46e34c4409 not found: ID does not exist" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.592420 4840 scope.go:117] "RemoveContainer" containerID="807dbc9df37357f57e1252dc8f83d1de62c3336b8cb4150169bcddd5321ae0e3" Sep 30 14:13:38 crc kubenswrapper[4840]: E0930 14:13:38.593242 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"807dbc9df37357f57e1252dc8f83d1de62c3336b8cb4150169bcddd5321ae0e3\": container with ID starting with 807dbc9df37357f57e1252dc8f83d1de62c3336b8cb4150169bcddd5321ae0e3 not found: ID does not exist" containerID="807dbc9df37357f57e1252dc8f83d1de62c3336b8cb4150169bcddd5321ae0e3" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.593282 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"807dbc9df37357f57e1252dc8f83d1de62c3336b8cb4150169bcddd5321ae0e3"} err="failed to get container status \"807dbc9df37357f57e1252dc8f83d1de62c3336b8cb4150169bcddd5321ae0e3\": rpc error: code = NotFound desc = could not find container \"807dbc9df37357f57e1252dc8f83d1de62c3336b8cb4150169bcddd5321ae0e3\": container with ID starting with 807dbc9df37357f57e1252dc8f83d1de62c3336b8cb4150169bcddd5321ae0e3 not found: ID does not exist" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.597719 4840 scope.go:117] "RemoveContainer" containerID="2edee90f84eb32c953eeb5dfa1fc2729836da354666c490c7c4bac46e34c4409" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.598359 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2edee90f84eb32c953eeb5dfa1fc2729836da354666c490c7c4bac46e34c4409"} err="failed to get container status \"2edee90f84eb32c953eeb5dfa1fc2729836da354666c490c7c4bac46e34c4409\": rpc error: code = NotFound desc = could not find container \"2edee90f84eb32c953eeb5dfa1fc2729836da354666c490c7c4bac46e34c4409\": container with ID starting with 2edee90f84eb32c953eeb5dfa1fc2729836da354666c490c7c4bac46e34c4409 not found: ID does not exist" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.598378 4840 scope.go:117] "RemoveContainer" containerID="807dbc9df37357f57e1252dc8f83d1de62c3336b8cb4150169bcddd5321ae0e3" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.598767 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"807dbc9df37357f57e1252dc8f83d1de62c3336b8cb4150169bcddd5321ae0e3"} err="failed to get container status \"807dbc9df37357f57e1252dc8f83d1de62c3336b8cb4150169bcddd5321ae0e3\": rpc error: code = NotFound desc = could not find container \"807dbc9df37357f57e1252dc8f83d1de62c3336b8cb4150169bcddd5321ae0e3\": container with ID starting with 807dbc9df37357f57e1252dc8f83d1de62c3336b8cb4150169bcddd5321ae0e3 not found: ID does not exist" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.598789 4840 scope.go:117] "RemoveContainer" containerID="71b465222c8ca2c7063b1632b64ef4d2b3dbe73248c1b00977324af0871011fc" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.782795 4840 scope.go:117] "RemoveContainer" containerID="91f9f1d766fa7013e2cf435e7a46c86870d1a985c1f7db12ac2763f63a8b3730" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.895325 4840 scope.go:117] "RemoveContainer" containerID="71b465222c8ca2c7063b1632b64ef4d2b3dbe73248c1b00977324af0871011fc" Sep 30 14:13:38 crc kubenswrapper[4840]: E0930 14:13:38.895838 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71b465222c8ca2c7063b1632b64ef4d2b3dbe73248c1b00977324af0871011fc\": container with ID starting with 71b465222c8ca2c7063b1632b64ef4d2b3dbe73248c1b00977324af0871011fc not found: ID does not exist" containerID="71b465222c8ca2c7063b1632b64ef4d2b3dbe73248c1b00977324af0871011fc" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.895896 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71b465222c8ca2c7063b1632b64ef4d2b3dbe73248c1b00977324af0871011fc"} err="failed to get container status \"71b465222c8ca2c7063b1632b64ef4d2b3dbe73248c1b00977324af0871011fc\": rpc error: code = NotFound desc = could not find container \"71b465222c8ca2c7063b1632b64ef4d2b3dbe73248c1b00977324af0871011fc\": container with ID starting with 71b465222c8ca2c7063b1632b64ef4d2b3dbe73248c1b00977324af0871011fc not found: ID does not exist" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.895929 4840 scope.go:117] "RemoveContainer" containerID="91f9f1d766fa7013e2cf435e7a46c86870d1a985c1f7db12ac2763f63a8b3730" Sep 30 14:13:38 crc kubenswrapper[4840]: E0930 14:13:38.896262 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91f9f1d766fa7013e2cf435e7a46c86870d1a985c1f7db12ac2763f63a8b3730\": container with ID starting with 91f9f1d766fa7013e2cf435e7a46c86870d1a985c1f7db12ac2763f63a8b3730 not found: ID does not exist" containerID="91f9f1d766fa7013e2cf435e7a46c86870d1a985c1f7db12ac2763f63a8b3730" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.896290 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91f9f1d766fa7013e2cf435e7a46c86870d1a985c1f7db12ac2763f63a8b3730"} err="failed to get container status \"91f9f1d766fa7013e2cf435e7a46c86870d1a985c1f7db12ac2763f63a8b3730\": rpc error: code = NotFound desc = could not find container \"91f9f1d766fa7013e2cf435e7a46c86870d1a985c1f7db12ac2763f63a8b3730\": container with ID starting with 91f9f1d766fa7013e2cf435e7a46c86870d1a985c1f7db12ac2763f63a8b3730 not found: ID does not exist" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.896307 4840 scope.go:117] "RemoveContainer" containerID="71b465222c8ca2c7063b1632b64ef4d2b3dbe73248c1b00977324af0871011fc" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.896687 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71b465222c8ca2c7063b1632b64ef4d2b3dbe73248c1b00977324af0871011fc"} err="failed to get container status \"71b465222c8ca2c7063b1632b64ef4d2b3dbe73248c1b00977324af0871011fc\": rpc error: code = NotFound desc = could not find container \"71b465222c8ca2c7063b1632b64ef4d2b3dbe73248c1b00977324af0871011fc\": container with ID starting with 71b465222c8ca2c7063b1632b64ef4d2b3dbe73248c1b00977324af0871011fc not found: ID does not exist" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.896715 4840 scope.go:117] "RemoveContainer" containerID="91f9f1d766fa7013e2cf435e7a46c86870d1a985c1f7db12ac2763f63a8b3730" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.897033 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91f9f1d766fa7013e2cf435e7a46c86870d1a985c1f7db12ac2763f63a8b3730"} err="failed to get container status \"91f9f1d766fa7013e2cf435e7a46c86870d1a985c1f7db12ac2763f63a8b3730\": rpc error: code = NotFound desc = could not find container \"91f9f1d766fa7013e2cf435e7a46c86870d1a985c1f7db12ac2763f63a8b3730\": container with ID starting with 91f9f1d766fa7013e2cf435e7a46c86870d1a985c1f7db12ac2763f63a8b3730 not found: ID does not exist" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.897090 4840 scope.go:117] "RemoveContainer" containerID="6d38e955431f3f93cdb79f42366f8efd916449357858ee1e88db0ae9f7808b40" Sep 30 14:13:38 crc kubenswrapper[4840]: I0930 14:13:38.963643 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5fbb8d5d4d-g2v89" Sep 30 14:13:39 crc kubenswrapper[4840]: I0930 14:13:39.191707 4840 scope.go:117] "RemoveContainer" containerID="1e4111d6b9033739a506b5a02fccfc6924f5a5fd075ba04c56d92cc8bc086c9e" Sep 30 14:13:39 crc kubenswrapper[4840]: I0930 14:13:39.234793 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9338f796-fcc1-4b71-8c3e-d48d5234d83d","Type":"ContainerStarted","Data":"9aa6a9fbf32866a7f01e8a877ea38fa3904444dda11461efa18674c9d1785f61"} Sep 30 14:13:39 crc kubenswrapper[4840]: I0930 14:13:39.238730 4840 scope.go:117] "RemoveContainer" containerID="6d38e955431f3f93cdb79f42366f8efd916449357858ee1e88db0ae9f7808b40" Sep 30 14:13:39 crc kubenswrapper[4840]: E0930 14:13:39.243717 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d38e955431f3f93cdb79f42366f8efd916449357858ee1e88db0ae9f7808b40\": container with ID starting with 6d38e955431f3f93cdb79f42366f8efd916449357858ee1e88db0ae9f7808b40 not found: ID does not exist" containerID="6d38e955431f3f93cdb79f42366f8efd916449357858ee1e88db0ae9f7808b40" Sep 30 14:13:39 crc kubenswrapper[4840]: I0930 14:13:39.243756 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d38e955431f3f93cdb79f42366f8efd916449357858ee1e88db0ae9f7808b40"} err="failed to get container status \"6d38e955431f3f93cdb79f42366f8efd916449357858ee1e88db0ae9f7808b40\": rpc error: code = NotFound desc = could not find container \"6d38e955431f3f93cdb79f42366f8efd916449357858ee1e88db0ae9f7808b40\": container with ID starting with 6d38e955431f3f93cdb79f42366f8efd916449357858ee1e88db0ae9f7808b40 not found: ID does not exist" Sep 30 14:13:39 crc kubenswrapper[4840]: I0930 14:13:39.243780 4840 scope.go:117] "RemoveContainer" containerID="1e4111d6b9033739a506b5a02fccfc6924f5a5fd075ba04c56d92cc8bc086c9e" Sep 30 14:13:39 crc kubenswrapper[4840]: E0930 14:13:39.252837 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e4111d6b9033739a506b5a02fccfc6924f5a5fd075ba04c56d92cc8bc086c9e\": container with ID starting with 1e4111d6b9033739a506b5a02fccfc6924f5a5fd075ba04c56d92cc8bc086c9e not found: ID does not exist" containerID="1e4111d6b9033739a506b5a02fccfc6924f5a5fd075ba04c56d92cc8bc086c9e" Sep 30 14:13:39 crc kubenswrapper[4840]: I0930 14:13:39.252870 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e4111d6b9033739a506b5a02fccfc6924f5a5fd075ba04c56d92cc8bc086c9e"} err="failed to get container status \"1e4111d6b9033739a506b5a02fccfc6924f5a5fd075ba04c56d92cc8bc086c9e\": rpc error: code = NotFound desc = could not find container \"1e4111d6b9033739a506b5a02fccfc6924f5a5fd075ba04c56d92cc8bc086c9e\": container with ID starting with 1e4111d6b9033739a506b5a02fccfc6924f5a5fd075ba04c56d92cc8bc086c9e not found: ID does not exist" Sep 30 14:13:39 crc kubenswrapper[4840]: I0930 14:13:39.252888 4840 scope.go:117] "RemoveContainer" containerID="6d38e955431f3f93cdb79f42366f8efd916449357858ee1e88db0ae9f7808b40" Sep 30 14:13:39 crc kubenswrapper[4840]: I0930 14:13:39.254564 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d38e955431f3f93cdb79f42366f8efd916449357858ee1e88db0ae9f7808b40"} err="failed to get container status \"6d38e955431f3f93cdb79f42366f8efd916449357858ee1e88db0ae9f7808b40\": rpc error: code = NotFound desc = could not find container \"6d38e955431f3f93cdb79f42366f8efd916449357858ee1e88db0ae9f7808b40\": container with ID starting with 6d38e955431f3f93cdb79f42366f8efd916449357858ee1e88db0ae9f7808b40 not found: ID does not exist" Sep 30 14:13:39 crc kubenswrapper[4840]: I0930 14:13:39.254588 4840 scope.go:117] "RemoveContainer" containerID="1e4111d6b9033739a506b5a02fccfc6924f5a5fd075ba04c56d92cc8bc086c9e" Sep 30 14:13:39 crc kubenswrapper[4840]: I0930 14:13:39.255844 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e4111d6b9033739a506b5a02fccfc6924f5a5fd075ba04c56d92cc8bc086c9e"} err="failed to get container status \"1e4111d6b9033739a506b5a02fccfc6924f5a5fd075ba04c56d92cc8bc086c9e\": rpc error: code = NotFound desc = could not find container \"1e4111d6b9033739a506b5a02fccfc6924f5a5fd075ba04c56d92cc8bc086c9e\": container with ID starting with 1e4111d6b9033739a506b5a02fccfc6924f5a5fd075ba04c56d92cc8bc086c9e not found: ID does not exist" Sep 30 14:13:39 crc kubenswrapper[4840]: I0930 14:13:39.259202 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"eb01e3a9-61de-4dc8-954b-221d80a675fb","Type":"ContainerStarted","Data":"4a9c74a00e7659de6b465e4126383d5f71b515f0ef65d73ffa337a73874dbdaa"} Sep 30 14:13:40 crc kubenswrapper[4840]: I0930 14:13:40.138852 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47067ffa-fc01-48b4-abcd-4405053eadca" path="/var/lib/kubelet/pods/47067ffa-fc01-48b4-abcd-4405053eadca/volumes" Sep 30 14:13:40 crc kubenswrapper[4840]: I0930 14:13:40.147254 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7f0b012-f515-434f-919e-5f0409c8e554" path="/var/lib/kubelet/pods/a7f0b012-f515-434f-919e-5f0409c8e554/volumes" Sep 30 14:13:40 crc kubenswrapper[4840]: I0930 14:13:40.148172 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e141523f-ce9b-4626-b155-b0a82a3f6dc6" path="/var/lib/kubelet/pods/e141523f-ce9b-4626-b155-b0a82a3f6dc6/volumes" Sep 30 14:13:40 crc kubenswrapper[4840]: I0930 14:13:40.283841 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"eb01e3a9-61de-4dc8-954b-221d80a675fb","Type":"ContainerStarted","Data":"c07ae91bc1f5962e684c4cd5440e4d614f5b74dc5e033dc4254464dc9e7a3d0f"} Sep 30 14:13:40 crc kubenswrapper[4840]: I0930 14:13:40.283934 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Sep 30 14:13:40 crc kubenswrapper[4840]: I0930 14:13:40.287366 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9338f796-fcc1-4b71-8c3e-d48d5234d83d","Type":"ContainerStarted","Data":"834fd6078b398ee02a6379918df8e22ceb2b913854de1e041615c348cdacb890"} Sep 30 14:13:40 crc kubenswrapper[4840]: I0930 14:13:40.306074 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.306048462 podStartE2EDuration="3.306048462s" podCreationTimestamp="2025-09-30 14:13:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:13:40.299013298 +0000 UTC m=+1048.928099731" watchObservedRunningTime="2025-09-30 14:13:40.306048462 +0000 UTC m=+1048.935134885" Sep 30 14:13:40 crc kubenswrapper[4840]: I0930 14:13:40.478096 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-59645bff48-kxnsb" Sep 30 14:13:40 crc kubenswrapper[4840]: I0930 14:13:40.704263 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-94dfcc48d-mlrjd" Sep 30 14:13:40 crc kubenswrapper[4840]: I0930 14:13:40.790478 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-59645bff48-kxnsb" Sep 30 14:13:40 crc kubenswrapper[4840]: I0930 14:13:40.893323 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-5595d79c68-dgsth" Sep 30 14:13:41 crc kubenswrapper[4840]: I0930 14:13:41.073774 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Sep 30 14:13:41 crc kubenswrapper[4840]: I0930 14:13:41.186041 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5784cf869f-qs9s6" Sep 30 14:13:41 crc kubenswrapper[4840]: I0930 14:13:41.210908 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-658bb978b7-lrh6p" Sep 30 14:13:41 crc kubenswrapper[4840]: I0930 14:13:41.277889 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-zvfqz"] Sep 30 14:13:41 crc kubenswrapper[4840]: I0930 14:13:41.278185 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-84b966f6c9-zvfqz" podUID="5c9be753-da64-4599-b5d1-baf7f9be58d4" containerName="dnsmasq-dns" containerID="cri-o://5af93f66b8fa19a435a55e198257289a46272a26e52aa7cc585e6a87d507e6ce" gracePeriod=10 Sep 30 14:13:41 crc kubenswrapper[4840]: I0930 14:13:41.317780 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5fbb8d5d4d-g2v89"] Sep 30 14:13:41 crc kubenswrapper[4840]: I0930 14:13:41.319757 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5fbb8d5d4d-g2v89" podUID="9a51d9a5-67ae-488d-a1e7-003d7db4267f" containerName="neutron-api" containerID="cri-o://889bc6c8cd3512a46aa4cd11e1a67d3387b504acec3381ff67d135c78f8e8ccb" gracePeriod=30 Sep 30 14:13:41 crc kubenswrapper[4840]: I0930 14:13:41.320031 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5fbb8d5d4d-g2v89" podUID="9a51d9a5-67ae-488d-a1e7-003d7db4267f" containerName="neutron-httpd" containerID="cri-o://a1c3a2c622f9191ac9262da20051252eb62e825475ca9959fec3292d9d52112d" gracePeriod=30 Sep 30 14:13:41 crc kubenswrapper[4840]: I0930 14:13:41.593049 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Sep 30 14:13:41 crc kubenswrapper[4840]: I0930 14:13:41.657290 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 30 14:13:41 crc kubenswrapper[4840]: I0930 14:13:41.942700 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-zvfqz" Sep 30 14:13:42 crc kubenswrapper[4840]: I0930 14:13:42.109934 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5c9be753-da64-4599-b5d1-baf7f9be58d4-ovsdbserver-sb\") pod \"5c9be753-da64-4599-b5d1-baf7f9be58d4\" (UID: \"5c9be753-da64-4599-b5d1-baf7f9be58d4\") " Sep 30 14:13:42 crc kubenswrapper[4840]: I0930 14:13:42.110024 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5c9be753-da64-4599-b5d1-baf7f9be58d4-dns-swift-storage-0\") pod \"5c9be753-da64-4599-b5d1-baf7f9be58d4\" (UID: \"5c9be753-da64-4599-b5d1-baf7f9be58d4\") " Sep 30 14:13:42 crc kubenswrapper[4840]: I0930 14:13:42.110601 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c9be753-da64-4599-b5d1-baf7f9be58d4-config\") pod \"5c9be753-da64-4599-b5d1-baf7f9be58d4\" (UID: \"5c9be753-da64-4599-b5d1-baf7f9be58d4\") " Sep 30 14:13:42 crc kubenswrapper[4840]: I0930 14:13:42.110695 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt6lb\" (UniqueName: \"kubernetes.io/projected/5c9be753-da64-4599-b5d1-baf7f9be58d4-kube-api-access-vt6lb\") pod \"5c9be753-da64-4599-b5d1-baf7f9be58d4\" (UID: \"5c9be753-da64-4599-b5d1-baf7f9be58d4\") " Sep 30 14:13:42 crc kubenswrapper[4840]: I0930 14:13:42.110884 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5c9be753-da64-4599-b5d1-baf7f9be58d4-ovsdbserver-nb\") pod \"5c9be753-da64-4599-b5d1-baf7f9be58d4\" (UID: \"5c9be753-da64-4599-b5d1-baf7f9be58d4\") " Sep 30 14:13:42 crc kubenswrapper[4840]: I0930 14:13:42.110978 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c9be753-da64-4599-b5d1-baf7f9be58d4-dns-svc\") pod \"5c9be753-da64-4599-b5d1-baf7f9be58d4\" (UID: \"5c9be753-da64-4599-b5d1-baf7f9be58d4\") " Sep 30 14:13:42 crc kubenswrapper[4840]: I0930 14:13:42.130425 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c9be753-da64-4599-b5d1-baf7f9be58d4-kube-api-access-vt6lb" (OuterVolumeSpecName: "kube-api-access-vt6lb") pod "5c9be753-da64-4599-b5d1-baf7f9be58d4" (UID: "5c9be753-da64-4599-b5d1-baf7f9be58d4"). InnerVolumeSpecName "kube-api-access-vt6lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:13:42 crc kubenswrapper[4840]: I0930 14:13:42.215922 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt6lb\" (UniqueName: \"kubernetes.io/projected/5c9be753-da64-4599-b5d1-baf7f9be58d4-kube-api-access-vt6lb\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:42 crc kubenswrapper[4840]: I0930 14:13:42.240269 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c9be753-da64-4599-b5d1-baf7f9be58d4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5c9be753-da64-4599-b5d1-baf7f9be58d4" (UID: "5c9be753-da64-4599-b5d1-baf7f9be58d4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:13:42 crc kubenswrapper[4840]: I0930 14:13:42.240931 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c9be753-da64-4599-b5d1-baf7f9be58d4-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5c9be753-da64-4599-b5d1-baf7f9be58d4" (UID: "5c9be753-da64-4599-b5d1-baf7f9be58d4"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:13:42 crc kubenswrapper[4840]: I0930 14:13:42.255264 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c9be753-da64-4599-b5d1-baf7f9be58d4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5c9be753-da64-4599-b5d1-baf7f9be58d4" (UID: "5c9be753-da64-4599-b5d1-baf7f9be58d4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:13:42 crc kubenswrapper[4840]: I0930 14:13:42.260943 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c9be753-da64-4599-b5d1-baf7f9be58d4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5c9be753-da64-4599-b5d1-baf7f9be58d4" (UID: "5c9be753-da64-4599-b5d1-baf7f9be58d4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:13:42 crc kubenswrapper[4840]: I0930 14:13:42.263321 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c9be753-da64-4599-b5d1-baf7f9be58d4-config" (OuterVolumeSpecName: "config") pod "5c9be753-da64-4599-b5d1-baf7f9be58d4" (UID: "5c9be753-da64-4599-b5d1-baf7f9be58d4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:13:42 crc kubenswrapper[4840]: I0930 14:13:42.317540 4840 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5c9be753-da64-4599-b5d1-baf7f9be58d4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:42 crc kubenswrapper[4840]: I0930 14:13:42.317589 4840 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c9be753-da64-4599-b5d1-baf7f9be58d4-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:42 crc kubenswrapper[4840]: I0930 14:13:42.317602 4840 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5c9be753-da64-4599-b5d1-baf7f9be58d4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:42 crc kubenswrapper[4840]: I0930 14:13:42.317614 4840 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5c9be753-da64-4599-b5d1-baf7f9be58d4-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:42 crc kubenswrapper[4840]: I0930 14:13:42.317625 4840 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c9be753-da64-4599-b5d1-baf7f9be58d4-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:42 crc kubenswrapper[4840]: I0930 14:13:42.339120 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9338f796-fcc1-4b71-8c3e-d48d5234d83d","Type":"ContainerStarted","Data":"d234dbb728a108e0737708d70abef8344965f9b6ad50d4052e6ac805d36e26da"} Sep 30 14:13:42 crc kubenswrapper[4840]: I0930 14:13:42.340095 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 30 14:13:42 crc kubenswrapper[4840]: I0930 14:13:42.353848 4840 generic.go:334] "Generic (PLEG): container finished" podID="5c9be753-da64-4599-b5d1-baf7f9be58d4" containerID="5af93f66b8fa19a435a55e198257289a46272a26e52aa7cc585e6a87d507e6ce" exitCode=0 Sep 30 14:13:42 crc kubenswrapper[4840]: I0930 14:13:42.353922 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-zvfqz" event={"ID":"5c9be753-da64-4599-b5d1-baf7f9be58d4","Type":"ContainerDied","Data":"5af93f66b8fa19a435a55e198257289a46272a26e52aa7cc585e6a87d507e6ce"} Sep 30 14:13:42 crc kubenswrapper[4840]: I0930 14:13:42.353947 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-zvfqz" event={"ID":"5c9be753-da64-4599-b5d1-baf7f9be58d4","Type":"ContainerDied","Data":"e5428d36f647cdf3937529312de47d20ee328949c756a09885e9431a5246bb6c"} Sep 30 14:13:42 crc kubenswrapper[4840]: I0930 14:13:42.353962 4840 scope.go:117] "RemoveContainer" containerID="5af93f66b8fa19a435a55e198257289a46272a26e52aa7cc585e6a87d507e6ce" Sep 30 14:13:42 crc kubenswrapper[4840]: I0930 14:13:42.354095 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-zvfqz" Sep 30 14:13:42 crc kubenswrapper[4840]: I0930 14:13:42.363164 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.056710944 podStartE2EDuration="6.363148376s" podCreationTimestamp="2025-09-30 14:13:36 +0000 UTC" firstStartedPulling="2025-09-30 14:13:37.007278938 +0000 UTC m=+1045.636365361" lastFinishedPulling="2025-09-30 14:13:41.31371637 +0000 UTC m=+1049.942802793" observedRunningTime="2025-09-30 14:13:42.362100859 +0000 UTC m=+1050.991187292" watchObservedRunningTime="2025-09-30 14:13:42.363148376 +0000 UTC m=+1050.992234799" Sep 30 14:13:42 crc kubenswrapper[4840]: I0930 14:13:42.366109 4840 generic.go:334] "Generic (PLEG): container finished" podID="9a51d9a5-67ae-488d-a1e7-003d7db4267f" containerID="a1c3a2c622f9191ac9262da20051252eb62e825475ca9959fec3292d9d52112d" exitCode=0 Sep 30 14:13:42 crc kubenswrapper[4840]: I0930 14:13:42.366518 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3" containerName="cinder-scheduler" containerID="cri-o://4c101ba43c368a4ff85926c05dc7c575d9680e0252a1474b752fee17d9cb5e1d" gracePeriod=30 Sep 30 14:13:42 crc kubenswrapper[4840]: I0930 14:13:42.366259 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5fbb8d5d4d-g2v89" event={"ID":"9a51d9a5-67ae-488d-a1e7-003d7db4267f","Type":"ContainerDied","Data":"a1c3a2c622f9191ac9262da20051252eb62e825475ca9959fec3292d9d52112d"} Sep 30 14:13:42 crc kubenswrapper[4840]: I0930 14:13:42.367123 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3" containerName="probe" containerID="cri-o://4f6c20bdac5506dafa0cb9b4c126a0ebee27fdd4eea09653d408e0ecf420fbbe" gracePeriod=30 Sep 30 14:13:42 crc kubenswrapper[4840]: I0930 14:13:42.406736 4840 scope.go:117] "RemoveContainer" containerID="b59efce57c2c6f84cdf6482bb6869209f49de47b514336e597e1121d41db727d" Sep 30 14:13:42 crc kubenswrapper[4840]: I0930 14:13:42.448705 4840 scope.go:117] "RemoveContainer" containerID="5af93f66b8fa19a435a55e198257289a46272a26e52aa7cc585e6a87d507e6ce" Sep 30 14:13:42 crc kubenswrapper[4840]: I0930 14:13:42.448810 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-zvfqz"] Sep 30 14:13:42 crc kubenswrapper[4840]: E0930 14:13:42.452327 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5af93f66b8fa19a435a55e198257289a46272a26e52aa7cc585e6a87d507e6ce\": container with ID starting with 5af93f66b8fa19a435a55e198257289a46272a26e52aa7cc585e6a87d507e6ce not found: ID does not exist" containerID="5af93f66b8fa19a435a55e198257289a46272a26e52aa7cc585e6a87d507e6ce" Sep 30 14:13:42 crc kubenswrapper[4840]: I0930 14:13:42.452373 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5af93f66b8fa19a435a55e198257289a46272a26e52aa7cc585e6a87d507e6ce"} err="failed to get container status \"5af93f66b8fa19a435a55e198257289a46272a26e52aa7cc585e6a87d507e6ce\": rpc error: code = NotFound desc = could not find container \"5af93f66b8fa19a435a55e198257289a46272a26e52aa7cc585e6a87d507e6ce\": container with ID starting with 5af93f66b8fa19a435a55e198257289a46272a26e52aa7cc585e6a87d507e6ce not found: ID does not exist" Sep 30 14:13:42 crc kubenswrapper[4840]: I0930 14:13:42.452402 4840 scope.go:117] "RemoveContainer" containerID="b59efce57c2c6f84cdf6482bb6869209f49de47b514336e597e1121d41db727d" Sep 30 14:13:42 crc kubenswrapper[4840]: E0930 14:13:42.454702 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b59efce57c2c6f84cdf6482bb6869209f49de47b514336e597e1121d41db727d\": container with ID starting with b59efce57c2c6f84cdf6482bb6869209f49de47b514336e597e1121d41db727d not found: ID does not exist" containerID="b59efce57c2c6f84cdf6482bb6869209f49de47b514336e597e1121d41db727d" Sep 30 14:13:42 crc kubenswrapper[4840]: I0930 14:13:42.454741 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b59efce57c2c6f84cdf6482bb6869209f49de47b514336e597e1121d41db727d"} err="failed to get container status \"b59efce57c2c6f84cdf6482bb6869209f49de47b514336e597e1121d41db727d\": rpc error: code = NotFound desc = could not find container \"b59efce57c2c6f84cdf6482bb6869209f49de47b514336e597e1121d41db727d\": container with ID starting with b59efce57c2c6f84cdf6482bb6869209f49de47b514336e597e1121d41db727d not found: ID does not exist" Sep 30 14:13:42 crc kubenswrapper[4840]: I0930 14:13:42.462329 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-zvfqz"] Sep 30 14:13:43 crc kubenswrapper[4840]: I0930 14:13:43.177181 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-94dfcc48d-mlrjd" Sep 30 14:13:43 crc kubenswrapper[4840]: I0930 14:13:43.246317 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5595d79c68-dgsth"] Sep 30 14:13:43 crc kubenswrapper[4840]: I0930 14:13:43.246538 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5595d79c68-dgsth" podUID="5e9b87b2-4836-4857-8514-b2dcfe4de1f5" containerName="horizon-log" containerID="cri-o://019b8bf8dc37878bad1354f6206205178cb6737805a6453683b754b0a592f5d1" gracePeriod=30 Sep 30 14:13:43 crc kubenswrapper[4840]: I0930 14:13:43.246920 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5595d79c68-dgsth" podUID="5e9b87b2-4836-4857-8514-b2dcfe4de1f5" containerName="horizon" containerID="cri-o://728e90849e942f659a1349a083627796a9122008fae88346ec1107d99eefbab1" gracePeriod=30 Sep 30 14:13:43 crc kubenswrapper[4840]: I0930 14:13:43.262143 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5595d79c68-dgsth" podUID="5e9b87b2-4836-4857-8514-b2dcfe4de1f5" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": EOF" Sep 30 14:13:43 crc kubenswrapper[4840]: I0930 14:13:43.413857 4840 generic.go:334] "Generic (PLEG): container finished" podID="9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3" containerID="4f6c20bdac5506dafa0cb9b4c126a0ebee27fdd4eea09653d408e0ecf420fbbe" exitCode=0 Sep 30 14:13:43 crc kubenswrapper[4840]: I0930 14:13:43.414396 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3","Type":"ContainerDied","Data":"4f6c20bdac5506dafa0cb9b4c126a0ebee27fdd4eea09653d408e0ecf420fbbe"} Sep 30 14:13:44 crc kubenswrapper[4840]: I0930 14:13:44.142084 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c9be753-da64-4599-b5d1-baf7f9be58d4" path="/var/lib/kubelet/pods/5c9be753-da64-4599-b5d1-baf7f9be58d4/volumes" Sep 30 14:13:44 crc kubenswrapper[4840]: I0930 14:13:44.231387 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5fbb8d5d4d-g2v89" Sep 30 14:13:44 crc kubenswrapper[4840]: I0930 14:13:44.353140 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zclvs\" (UniqueName: \"kubernetes.io/projected/9a51d9a5-67ae-488d-a1e7-003d7db4267f-kube-api-access-zclvs\") pod \"9a51d9a5-67ae-488d-a1e7-003d7db4267f\" (UID: \"9a51d9a5-67ae-488d-a1e7-003d7db4267f\") " Sep 30 14:13:44 crc kubenswrapper[4840]: I0930 14:13:44.353247 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9a51d9a5-67ae-488d-a1e7-003d7db4267f-httpd-config\") pod \"9a51d9a5-67ae-488d-a1e7-003d7db4267f\" (UID: \"9a51d9a5-67ae-488d-a1e7-003d7db4267f\") " Sep 30 14:13:44 crc kubenswrapper[4840]: I0930 14:13:44.353294 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a51d9a5-67ae-488d-a1e7-003d7db4267f-combined-ca-bundle\") pod \"9a51d9a5-67ae-488d-a1e7-003d7db4267f\" (UID: \"9a51d9a5-67ae-488d-a1e7-003d7db4267f\") " Sep 30 14:13:44 crc kubenswrapper[4840]: I0930 14:13:44.353382 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a51d9a5-67ae-488d-a1e7-003d7db4267f-ovndb-tls-certs\") pod \"9a51d9a5-67ae-488d-a1e7-003d7db4267f\" (UID: \"9a51d9a5-67ae-488d-a1e7-003d7db4267f\") " Sep 30 14:13:44 crc kubenswrapper[4840]: I0930 14:13:44.353511 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9a51d9a5-67ae-488d-a1e7-003d7db4267f-config\") pod \"9a51d9a5-67ae-488d-a1e7-003d7db4267f\" (UID: \"9a51d9a5-67ae-488d-a1e7-003d7db4267f\") " Sep 30 14:13:44 crc kubenswrapper[4840]: I0930 14:13:44.359233 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a51d9a5-67ae-488d-a1e7-003d7db4267f-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "9a51d9a5-67ae-488d-a1e7-003d7db4267f" (UID: "9a51d9a5-67ae-488d-a1e7-003d7db4267f"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:44 crc kubenswrapper[4840]: I0930 14:13:44.365876 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a51d9a5-67ae-488d-a1e7-003d7db4267f-kube-api-access-zclvs" (OuterVolumeSpecName: "kube-api-access-zclvs") pod "9a51d9a5-67ae-488d-a1e7-003d7db4267f" (UID: "9a51d9a5-67ae-488d-a1e7-003d7db4267f"). InnerVolumeSpecName "kube-api-access-zclvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:13:44 crc kubenswrapper[4840]: I0930 14:13:44.409757 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a51d9a5-67ae-488d-a1e7-003d7db4267f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9a51d9a5-67ae-488d-a1e7-003d7db4267f" (UID: "9a51d9a5-67ae-488d-a1e7-003d7db4267f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:44 crc kubenswrapper[4840]: I0930 14:13:44.411743 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a51d9a5-67ae-488d-a1e7-003d7db4267f-config" (OuterVolumeSpecName: "config") pod "9a51d9a5-67ae-488d-a1e7-003d7db4267f" (UID: "9a51d9a5-67ae-488d-a1e7-003d7db4267f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:44 crc kubenswrapper[4840]: I0930 14:13:44.435495 4840 generic.go:334] "Generic (PLEG): container finished" podID="9a51d9a5-67ae-488d-a1e7-003d7db4267f" containerID="889bc6c8cd3512a46aa4cd11e1a67d3387b504acec3381ff67d135c78f8e8ccb" exitCode=0 Sep 30 14:13:44 crc kubenswrapper[4840]: I0930 14:13:44.436535 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5fbb8d5d4d-g2v89" Sep 30 14:13:44 crc kubenswrapper[4840]: I0930 14:13:44.436999 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5fbb8d5d4d-g2v89" event={"ID":"9a51d9a5-67ae-488d-a1e7-003d7db4267f","Type":"ContainerDied","Data":"889bc6c8cd3512a46aa4cd11e1a67d3387b504acec3381ff67d135c78f8e8ccb"} Sep 30 14:13:44 crc kubenswrapper[4840]: I0930 14:13:44.437029 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5fbb8d5d4d-g2v89" event={"ID":"9a51d9a5-67ae-488d-a1e7-003d7db4267f","Type":"ContainerDied","Data":"058945b169bf7708ddcedb1ca2e22711684ab3b132c19cf217a8cab6341ef1d9"} Sep 30 14:13:44 crc kubenswrapper[4840]: I0930 14:13:44.437044 4840 scope.go:117] "RemoveContainer" containerID="a1c3a2c622f9191ac9262da20051252eb62e825475ca9959fec3292d9d52112d" Sep 30 14:13:44 crc kubenswrapper[4840]: I0930 14:13:44.455461 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zclvs\" (UniqueName: \"kubernetes.io/projected/9a51d9a5-67ae-488d-a1e7-003d7db4267f-kube-api-access-zclvs\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:44 crc kubenswrapper[4840]: I0930 14:13:44.455805 4840 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9a51d9a5-67ae-488d-a1e7-003d7db4267f-httpd-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:44 crc kubenswrapper[4840]: I0930 14:13:44.455816 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a51d9a5-67ae-488d-a1e7-003d7db4267f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:44 crc kubenswrapper[4840]: I0930 14:13:44.455824 4840 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/9a51d9a5-67ae-488d-a1e7-003d7db4267f-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:44 crc kubenswrapper[4840]: I0930 14:13:44.456650 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a51d9a5-67ae-488d-a1e7-003d7db4267f-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "9a51d9a5-67ae-488d-a1e7-003d7db4267f" (UID: "9a51d9a5-67ae-488d-a1e7-003d7db4267f"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:44 crc kubenswrapper[4840]: I0930 14:13:44.469445 4840 scope.go:117] "RemoveContainer" containerID="889bc6c8cd3512a46aa4cd11e1a67d3387b504acec3381ff67d135c78f8e8ccb" Sep 30 14:13:44 crc kubenswrapper[4840]: I0930 14:13:44.494257 4840 scope.go:117] "RemoveContainer" containerID="a1c3a2c622f9191ac9262da20051252eb62e825475ca9959fec3292d9d52112d" Sep 30 14:13:44 crc kubenswrapper[4840]: E0930 14:13:44.494965 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1c3a2c622f9191ac9262da20051252eb62e825475ca9959fec3292d9d52112d\": container with ID starting with a1c3a2c622f9191ac9262da20051252eb62e825475ca9959fec3292d9d52112d not found: ID does not exist" containerID="a1c3a2c622f9191ac9262da20051252eb62e825475ca9959fec3292d9d52112d" Sep 30 14:13:44 crc kubenswrapper[4840]: I0930 14:13:44.495006 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1c3a2c622f9191ac9262da20051252eb62e825475ca9959fec3292d9d52112d"} err="failed to get container status \"a1c3a2c622f9191ac9262da20051252eb62e825475ca9959fec3292d9d52112d\": rpc error: code = NotFound desc = could not find container \"a1c3a2c622f9191ac9262da20051252eb62e825475ca9959fec3292d9d52112d\": container with ID starting with a1c3a2c622f9191ac9262da20051252eb62e825475ca9959fec3292d9d52112d not found: ID does not exist" Sep 30 14:13:44 crc kubenswrapper[4840]: I0930 14:13:44.495032 4840 scope.go:117] "RemoveContainer" containerID="889bc6c8cd3512a46aa4cd11e1a67d3387b504acec3381ff67d135c78f8e8ccb" Sep 30 14:13:44 crc kubenswrapper[4840]: E0930 14:13:44.495337 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"889bc6c8cd3512a46aa4cd11e1a67d3387b504acec3381ff67d135c78f8e8ccb\": container with ID starting with 889bc6c8cd3512a46aa4cd11e1a67d3387b504acec3381ff67d135c78f8e8ccb not found: ID does not exist" containerID="889bc6c8cd3512a46aa4cd11e1a67d3387b504acec3381ff67d135c78f8e8ccb" Sep 30 14:13:44 crc kubenswrapper[4840]: I0930 14:13:44.495371 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"889bc6c8cd3512a46aa4cd11e1a67d3387b504acec3381ff67d135c78f8e8ccb"} err="failed to get container status \"889bc6c8cd3512a46aa4cd11e1a67d3387b504acec3381ff67d135c78f8e8ccb\": rpc error: code = NotFound desc = could not find container \"889bc6c8cd3512a46aa4cd11e1a67d3387b504acec3381ff67d135c78f8e8ccb\": container with ID starting with 889bc6c8cd3512a46aa4cd11e1a67d3387b504acec3381ff67d135c78f8e8ccb not found: ID does not exist" Sep 30 14:13:44 crc kubenswrapper[4840]: I0930 14:13:44.557455 4840 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a51d9a5-67ae-488d-a1e7-003d7db4267f-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:44 crc kubenswrapper[4840]: I0930 14:13:44.797729 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5fbb8d5d4d-g2v89"] Sep 30 14:13:44 crc kubenswrapper[4840]: I0930 14:13:44.807152 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5fbb8d5d4d-g2v89"] Sep 30 14:13:45 crc kubenswrapper[4840]: I0930 14:13:45.235530 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-5595d79c68-dgsth" Sep 30 14:13:46 crc kubenswrapper[4840]: I0930 14:13:46.128474 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a51d9a5-67ae-488d-a1e7-003d7db4267f" path="/var/lib/kubelet/pods/9a51d9a5-67ae-488d-a1e7-003d7db4267f/volumes" Sep 30 14:13:46 crc kubenswrapper[4840]: I0930 14:13:46.457001 4840 generic.go:334] "Generic (PLEG): container finished" podID="9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3" containerID="4c101ba43c368a4ff85926c05dc7c575d9680e0252a1474b752fee17d9cb5e1d" exitCode=0 Sep 30 14:13:46 crc kubenswrapper[4840]: I0930 14:13:46.457067 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3","Type":"ContainerDied","Data":"4c101ba43c368a4ff85926c05dc7c575d9680e0252a1474b752fee17d9cb5e1d"} Sep 30 14:13:46 crc kubenswrapper[4840]: I0930 14:13:46.907067 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-784f485bd4-gztrl" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.070518 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.150456 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-784f485bd4-gztrl" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.203887 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3-etc-machine-id\") pod \"9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3\" (UID: \"9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3\") " Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.203959 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3-combined-ca-bundle\") pod \"9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3\" (UID: \"9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3\") " Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.204034 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3-config-data\") pod \"9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3\" (UID: \"9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3\") " Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.204008 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3" (UID: "9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.204081 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2qrqw\" (UniqueName: \"kubernetes.io/projected/9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3-kube-api-access-2qrqw\") pod \"9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3\" (UID: \"9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3\") " Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.204176 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3-scripts\") pod \"9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3\" (UID: \"9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3\") " Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.204193 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3-config-data-custom\") pod \"9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3\" (UID: \"9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3\") " Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.204626 4840 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3-etc-machine-id\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.214700 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3" (UID: "9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.214873 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3-kube-api-access-2qrqw" (OuterVolumeSpecName: "kube-api-access-2qrqw") pod "9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3" (UID: "9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3"). InnerVolumeSpecName "kube-api-access-2qrqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.225474 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-59645bff48-kxnsb"] Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.225854 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-59645bff48-kxnsb" podUID="1650244d-222b-4c4e-80f8-720393e7eb67" containerName="barbican-api-log" containerID="cri-o://af005a4cb3f5c14bbf6ea4d65b23bee0a91d02cf2ba5b71bf0fd3c936f7904a4" gracePeriod=30 Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.226068 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-59645bff48-kxnsb" podUID="1650244d-222b-4c4e-80f8-720393e7eb67" containerName="barbican-api" containerID="cri-o://9bbeaf4ecaa27d44f6cc27c9d44e044a6e6f58d68591aa17c80309694e88e34e" gracePeriod=30 Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.226955 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3-scripts" (OuterVolumeSpecName: "scripts") pod "9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3" (UID: "9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.312736 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2qrqw\" (UniqueName: \"kubernetes.io/projected/9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3-kube-api-access-2qrqw\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.312767 4840 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.312776 4840 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3-config-data-custom\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.318002 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3" (UID: "9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.415651 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.418154 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3-config-data" (OuterVolumeSpecName: "config-data") pod "9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3" (UID: "9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.466712 4840 generic.go:334] "Generic (PLEG): container finished" podID="1650244d-222b-4c4e-80f8-720393e7eb67" containerID="af005a4cb3f5c14bbf6ea4d65b23bee0a91d02cf2ba5b71bf0fd3c936f7904a4" exitCode=143 Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.466752 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-59645bff48-kxnsb" event={"ID":"1650244d-222b-4c4e-80f8-720393e7eb67","Type":"ContainerDied","Data":"af005a4cb3f5c14bbf6ea4d65b23bee0a91d02cf2ba5b71bf0fd3c936f7904a4"} Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.469128 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3","Type":"ContainerDied","Data":"861cc27309a90677d4bc9cd4e3e49f491810252bcd526bcdb2faea77f3b9fea5"} Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.469167 4840 scope.go:117] "RemoveContainer" containerID="4f6c20bdac5506dafa0cb9b4c126a0ebee27fdd4eea09653d408e0ecf420fbbe" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.469203 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.508141 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.513737 4840 scope.go:117] "RemoveContainer" containerID="4c101ba43c368a4ff85926c05dc7c575d9680e0252a1474b752fee17d9cb5e1d" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.517850 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.526779 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.536641 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Sep 30 14:13:47 crc kubenswrapper[4840]: E0930 14:13:47.537039 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47067ffa-fc01-48b4-abcd-4405053eadca" containerName="horizon-log" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.537054 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="47067ffa-fc01-48b4-abcd-4405053eadca" containerName="horizon-log" Sep 30 14:13:47 crc kubenswrapper[4840]: E0930 14:13:47.537070 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47067ffa-fc01-48b4-abcd-4405053eadca" containerName="horizon" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.537077 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="47067ffa-fc01-48b4-abcd-4405053eadca" containerName="horizon" Sep 30 14:13:47 crc kubenswrapper[4840]: E0930 14:13:47.537090 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c9be753-da64-4599-b5d1-baf7f9be58d4" containerName="init" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.537098 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c9be753-da64-4599-b5d1-baf7f9be58d4" containerName="init" Sep 30 14:13:47 crc kubenswrapper[4840]: E0930 14:13:47.537107 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c9be753-da64-4599-b5d1-baf7f9be58d4" containerName="dnsmasq-dns" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.537116 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c9be753-da64-4599-b5d1-baf7f9be58d4" containerName="dnsmasq-dns" Sep 30 14:13:47 crc kubenswrapper[4840]: E0930 14:13:47.537131 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a51d9a5-67ae-488d-a1e7-003d7db4267f" containerName="neutron-httpd" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.537138 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a51d9a5-67ae-488d-a1e7-003d7db4267f" containerName="neutron-httpd" Sep 30 14:13:47 crc kubenswrapper[4840]: E0930 14:13:47.537158 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a51d9a5-67ae-488d-a1e7-003d7db4267f" containerName="neutron-api" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.537165 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a51d9a5-67ae-488d-a1e7-003d7db4267f" containerName="neutron-api" Sep 30 14:13:47 crc kubenswrapper[4840]: E0930 14:13:47.537183 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e141523f-ce9b-4626-b155-b0a82a3f6dc6" containerName="horizon" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.537192 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="e141523f-ce9b-4626-b155-b0a82a3f6dc6" containerName="horizon" Sep 30 14:13:47 crc kubenswrapper[4840]: E0930 14:13:47.537204 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e141523f-ce9b-4626-b155-b0a82a3f6dc6" containerName="horizon-log" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.537212 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="e141523f-ce9b-4626-b155-b0a82a3f6dc6" containerName="horizon-log" Sep 30 14:13:47 crc kubenswrapper[4840]: E0930 14:13:47.537233 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7f0b012-f515-434f-919e-5f0409c8e554" containerName="horizon" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.537240 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7f0b012-f515-434f-919e-5f0409c8e554" containerName="horizon" Sep 30 14:13:47 crc kubenswrapper[4840]: E0930 14:13:47.537254 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3" containerName="cinder-scheduler" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.537262 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3" containerName="cinder-scheduler" Sep 30 14:13:47 crc kubenswrapper[4840]: E0930 14:13:47.537275 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3" containerName="probe" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.537282 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3" containerName="probe" Sep 30 14:13:47 crc kubenswrapper[4840]: E0930 14:13:47.537298 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7f0b012-f515-434f-919e-5f0409c8e554" containerName="horizon-log" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.537306 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7f0b012-f515-434f-919e-5f0409c8e554" containerName="horizon-log" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.537816 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="e141523f-ce9b-4626-b155-b0a82a3f6dc6" containerName="horizon" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.537840 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7f0b012-f515-434f-919e-5f0409c8e554" containerName="horizon" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.537854 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="e141523f-ce9b-4626-b155-b0a82a3f6dc6" containerName="horizon-log" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.537869 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3" containerName="cinder-scheduler" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.537886 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a51d9a5-67ae-488d-a1e7-003d7db4267f" containerName="neutron-httpd" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.537893 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c9be753-da64-4599-b5d1-baf7f9be58d4" containerName="dnsmasq-dns" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.537906 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3" containerName="probe" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.537916 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="47067ffa-fc01-48b4-abcd-4405053eadca" containerName="horizon" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.537932 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a51d9a5-67ae-488d-a1e7-003d7db4267f" containerName="neutron-api" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.537947 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7f0b012-f515-434f-919e-5f0409c8e554" containerName="horizon-log" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.537957 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="47067ffa-fc01-48b4-abcd-4405053eadca" containerName="horizon-log" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.540095 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.541871 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.546935 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.619278 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvpsb\" (UniqueName: \"kubernetes.io/projected/01feefa5-d23e-4189-9d57-84bae2e247e2-kube-api-access-bvpsb\") pod \"cinder-scheduler-0\" (UID: \"01feefa5-d23e-4189-9d57-84bae2e247e2\") " pod="openstack/cinder-scheduler-0" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.619392 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/01feefa5-d23e-4189-9d57-84bae2e247e2-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"01feefa5-d23e-4189-9d57-84bae2e247e2\") " pod="openstack/cinder-scheduler-0" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.619604 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/01feefa5-d23e-4189-9d57-84bae2e247e2-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"01feefa5-d23e-4189-9d57-84bae2e247e2\") " pod="openstack/cinder-scheduler-0" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.619704 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01feefa5-d23e-4189-9d57-84bae2e247e2-config-data\") pod \"cinder-scheduler-0\" (UID: \"01feefa5-d23e-4189-9d57-84bae2e247e2\") " pod="openstack/cinder-scheduler-0" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.619728 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01feefa5-d23e-4189-9d57-84bae2e247e2-scripts\") pod \"cinder-scheduler-0\" (UID: \"01feefa5-d23e-4189-9d57-84bae2e247e2\") " pod="openstack/cinder-scheduler-0" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.619787 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01feefa5-d23e-4189-9d57-84bae2e247e2-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"01feefa5-d23e-4189-9d57-84bae2e247e2\") " pod="openstack/cinder-scheduler-0" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.720909 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/01feefa5-d23e-4189-9d57-84bae2e247e2-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"01feefa5-d23e-4189-9d57-84bae2e247e2\") " pod="openstack/cinder-scheduler-0" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.721175 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/01feefa5-d23e-4189-9d57-84bae2e247e2-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"01feefa5-d23e-4189-9d57-84bae2e247e2\") " pod="openstack/cinder-scheduler-0" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.721331 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01feefa5-d23e-4189-9d57-84bae2e247e2-config-data\") pod \"cinder-scheduler-0\" (UID: \"01feefa5-d23e-4189-9d57-84bae2e247e2\") " pod="openstack/cinder-scheduler-0" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.721418 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01feefa5-d23e-4189-9d57-84bae2e247e2-scripts\") pod \"cinder-scheduler-0\" (UID: \"01feefa5-d23e-4189-9d57-84bae2e247e2\") " pod="openstack/cinder-scheduler-0" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.721527 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01feefa5-d23e-4189-9d57-84bae2e247e2-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"01feefa5-d23e-4189-9d57-84bae2e247e2\") " pod="openstack/cinder-scheduler-0" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.721702 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvpsb\" (UniqueName: \"kubernetes.io/projected/01feefa5-d23e-4189-9d57-84bae2e247e2-kube-api-access-bvpsb\") pod \"cinder-scheduler-0\" (UID: \"01feefa5-d23e-4189-9d57-84bae2e247e2\") " pod="openstack/cinder-scheduler-0" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.722343 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/01feefa5-d23e-4189-9d57-84bae2e247e2-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"01feefa5-d23e-4189-9d57-84bae2e247e2\") " pod="openstack/cinder-scheduler-0" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.726262 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/01feefa5-d23e-4189-9d57-84bae2e247e2-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"01feefa5-d23e-4189-9d57-84bae2e247e2\") " pod="openstack/cinder-scheduler-0" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.727282 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01feefa5-d23e-4189-9d57-84bae2e247e2-scripts\") pod \"cinder-scheduler-0\" (UID: \"01feefa5-d23e-4189-9d57-84bae2e247e2\") " pod="openstack/cinder-scheduler-0" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.731983 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01feefa5-d23e-4189-9d57-84bae2e247e2-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"01feefa5-d23e-4189-9d57-84bae2e247e2\") " pod="openstack/cinder-scheduler-0" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.741599 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01feefa5-d23e-4189-9d57-84bae2e247e2-config-data\") pod \"cinder-scheduler-0\" (UID: \"01feefa5-d23e-4189-9d57-84bae2e247e2\") " pod="openstack/cinder-scheduler-0" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.758103 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvpsb\" (UniqueName: \"kubernetes.io/projected/01feefa5-d23e-4189-9d57-84bae2e247e2-kube-api-access-bvpsb\") pod \"cinder-scheduler-0\" (UID: \"01feefa5-d23e-4189-9d57-84bae2e247e2\") " pod="openstack/cinder-scheduler-0" Sep 30 14:13:47 crc kubenswrapper[4840]: I0930 14:13:47.877220 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 30 14:13:48 crc kubenswrapper[4840]: I0930 14:13:48.126861 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3" path="/var/lib/kubelet/pods/9bdcbed9-fac8-4c80-ae2e-3570c9f2f6d3/volumes" Sep 30 14:13:48 crc kubenswrapper[4840]: I0930 14:13:48.420491 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 30 14:13:48 crc kubenswrapper[4840]: I0930 14:13:48.479203 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"01feefa5-d23e-4189-9d57-84bae2e247e2","Type":"ContainerStarted","Data":"4cd65500494df5cbc1e762f6d45302c45fc790c44bd23ada61ecb7f6315d0598"} Sep 30 14:13:48 crc kubenswrapper[4840]: I0930 14:13:48.518247 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-6c78d49dc8-kqrxd" Sep 30 14:13:48 crc kubenswrapper[4840]: I0930 14:13:48.681281 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5595d79c68-dgsth" podUID="5e9b87b2-4836-4857-8514-b2dcfe4de1f5" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": read tcp 10.217.0.2:53824->10.217.0.149:8443: read: connection reset by peer" Sep 30 14:13:49 crc kubenswrapper[4840]: I0930 14:13:49.496520 4840 generic.go:334] "Generic (PLEG): container finished" podID="5e9b87b2-4836-4857-8514-b2dcfe4de1f5" containerID="728e90849e942f659a1349a083627796a9122008fae88346ec1107d99eefbab1" exitCode=0 Sep 30 14:13:49 crc kubenswrapper[4840]: I0930 14:13:49.497092 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5595d79c68-dgsth" event={"ID":"5e9b87b2-4836-4857-8514-b2dcfe4de1f5","Type":"ContainerDied","Data":"728e90849e942f659a1349a083627796a9122008fae88346ec1107d99eefbab1"} Sep 30 14:13:49 crc kubenswrapper[4840]: I0930 14:13:49.499833 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"01feefa5-d23e-4189-9d57-84bae2e247e2","Type":"ContainerStarted","Data":"fdab9837a052e6adc4f13b97bbb76078e338f31639619bd834811deed6cb2a06"} Sep 30 14:13:49 crc kubenswrapper[4840]: I0930 14:13:49.636461 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Sep 30 14:13:50 crc kubenswrapper[4840]: I0930 14:13:50.416695 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-59645bff48-kxnsb" podUID="1650244d-222b-4c4e-80f8-720393e7eb67" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.162:9311/healthcheck\": read tcp 10.217.0.2:55806->10.217.0.162:9311: read: connection reset by peer" Sep 30 14:13:50 crc kubenswrapper[4840]: I0930 14:13:50.416792 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-59645bff48-kxnsb" podUID="1650244d-222b-4c4e-80f8-720393e7eb67" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.162:9311/healthcheck\": read tcp 10.217.0.2:55794->10.217.0.162:9311: read: connection reset by peer" Sep 30 14:13:50 crc kubenswrapper[4840]: I0930 14:13:50.526714 4840 generic.go:334] "Generic (PLEG): container finished" podID="1650244d-222b-4c4e-80f8-720393e7eb67" containerID="9bbeaf4ecaa27d44f6cc27c9d44e044a6e6f58d68591aa17c80309694e88e34e" exitCode=0 Sep 30 14:13:50 crc kubenswrapper[4840]: I0930 14:13:50.526788 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-59645bff48-kxnsb" event={"ID":"1650244d-222b-4c4e-80f8-720393e7eb67","Type":"ContainerDied","Data":"9bbeaf4ecaa27d44f6cc27c9d44e044a6e6f58d68591aa17c80309694e88e34e"} Sep 30 14:13:50 crc kubenswrapper[4840]: I0930 14:13:50.529697 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"01feefa5-d23e-4189-9d57-84bae2e247e2","Type":"ContainerStarted","Data":"3d08f29e9310d6a3d425c189bec5679db596911d2b26e0f126acf97f3e1685d1"} Sep 30 14:13:50 crc kubenswrapper[4840]: I0930 14:13:50.556010 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.5559782589999998 podStartE2EDuration="3.555978259s" podCreationTimestamp="2025-09-30 14:13:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:13:50.550038494 +0000 UTC m=+1059.179124917" watchObservedRunningTime="2025-09-30 14:13:50.555978259 +0000 UTC m=+1059.185064682" Sep 30 14:13:50 crc kubenswrapper[4840]: I0930 14:13:50.650411 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Sep 30 14:13:50 crc kubenswrapper[4840]: I0930 14:13:50.651729 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Sep 30 14:13:50 crc kubenswrapper[4840]: I0930 14:13:50.661326 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Sep 30 14:13:50 crc kubenswrapper[4840]: I0930 14:13:50.661336 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Sep 30 14:13:50 crc kubenswrapper[4840]: I0930 14:13:50.661817 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-49kqm" Sep 30 14:13:50 crc kubenswrapper[4840]: I0930 14:13:50.669897 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Sep 30 14:13:50 crc kubenswrapper[4840]: I0930 14:13:50.789895 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e234012b-317e-4c79-81a3-3991e1296058-combined-ca-bundle\") pod \"openstackclient\" (UID: \"e234012b-317e-4c79-81a3-3991e1296058\") " pod="openstack/openstackclient" Sep 30 14:13:50 crc kubenswrapper[4840]: I0930 14:13:50.789977 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9sjqz\" (UniqueName: \"kubernetes.io/projected/e234012b-317e-4c79-81a3-3991e1296058-kube-api-access-9sjqz\") pod \"openstackclient\" (UID: \"e234012b-317e-4c79-81a3-3991e1296058\") " pod="openstack/openstackclient" Sep 30 14:13:50 crc kubenswrapper[4840]: I0930 14:13:50.790044 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e234012b-317e-4c79-81a3-3991e1296058-openstack-config-secret\") pod \"openstackclient\" (UID: \"e234012b-317e-4c79-81a3-3991e1296058\") " pod="openstack/openstackclient" Sep 30 14:13:50 crc kubenswrapper[4840]: I0930 14:13:50.790121 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e234012b-317e-4c79-81a3-3991e1296058-openstack-config\") pod \"openstackclient\" (UID: \"e234012b-317e-4c79-81a3-3991e1296058\") " pod="openstack/openstackclient" Sep 30 14:13:50 crc kubenswrapper[4840]: I0930 14:13:50.892346 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e234012b-317e-4c79-81a3-3991e1296058-openstack-config\") pod \"openstackclient\" (UID: \"e234012b-317e-4c79-81a3-3991e1296058\") " pod="openstack/openstackclient" Sep 30 14:13:50 crc kubenswrapper[4840]: I0930 14:13:50.892457 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e234012b-317e-4c79-81a3-3991e1296058-combined-ca-bundle\") pod \"openstackclient\" (UID: \"e234012b-317e-4c79-81a3-3991e1296058\") " pod="openstack/openstackclient" Sep 30 14:13:50 crc kubenswrapper[4840]: I0930 14:13:50.892507 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9sjqz\" (UniqueName: \"kubernetes.io/projected/e234012b-317e-4c79-81a3-3991e1296058-kube-api-access-9sjqz\") pod \"openstackclient\" (UID: \"e234012b-317e-4c79-81a3-3991e1296058\") " pod="openstack/openstackclient" Sep 30 14:13:50 crc kubenswrapper[4840]: I0930 14:13:50.892572 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e234012b-317e-4c79-81a3-3991e1296058-openstack-config-secret\") pod \"openstackclient\" (UID: \"e234012b-317e-4c79-81a3-3991e1296058\") " pod="openstack/openstackclient" Sep 30 14:13:50 crc kubenswrapper[4840]: I0930 14:13:50.893839 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e234012b-317e-4c79-81a3-3991e1296058-openstack-config\") pod \"openstackclient\" (UID: \"e234012b-317e-4c79-81a3-3991e1296058\") " pod="openstack/openstackclient" Sep 30 14:13:50 crc kubenswrapper[4840]: I0930 14:13:50.900201 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e234012b-317e-4c79-81a3-3991e1296058-combined-ca-bundle\") pod \"openstackclient\" (UID: \"e234012b-317e-4c79-81a3-3991e1296058\") " pod="openstack/openstackclient" Sep 30 14:13:50 crc kubenswrapper[4840]: I0930 14:13:50.900495 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e234012b-317e-4c79-81a3-3991e1296058-openstack-config-secret\") pod \"openstackclient\" (UID: \"e234012b-317e-4c79-81a3-3991e1296058\") " pod="openstack/openstackclient" Sep 30 14:13:50 crc kubenswrapper[4840]: I0930 14:13:50.914221 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9sjqz\" (UniqueName: \"kubernetes.io/projected/e234012b-317e-4c79-81a3-3991e1296058-kube-api-access-9sjqz\") pod \"openstackclient\" (UID: \"e234012b-317e-4c79-81a3-3991e1296058\") " pod="openstack/openstackclient" Sep 30 14:13:50 crc kubenswrapper[4840]: I0930 14:13:50.980127 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Sep 30 14:13:51 crc kubenswrapper[4840]: I0930 14:13:51.014881 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-59645bff48-kxnsb" Sep 30 14:13:51 crc kubenswrapper[4840]: I0930 14:13:51.098588 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1650244d-222b-4c4e-80f8-720393e7eb67-logs\") pod \"1650244d-222b-4c4e-80f8-720393e7eb67\" (UID: \"1650244d-222b-4c4e-80f8-720393e7eb67\") " Sep 30 14:13:51 crc kubenswrapper[4840]: I0930 14:13:51.098860 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1650244d-222b-4c4e-80f8-720393e7eb67-config-data\") pod \"1650244d-222b-4c4e-80f8-720393e7eb67\" (UID: \"1650244d-222b-4c4e-80f8-720393e7eb67\") " Sep 30 14:13:51 crc kubenswrapper[4840]: I0930 14:13:51.098889 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1650244d-222b-4c4e-80f8-720393e7eb67-combined-ca-bundle\") pod \"1650244d-222b-4c4e-80f8-720393e7eb67\" (UID: \"1650244d-222b-4c4e-80f8-720393e7eb67\") " Sep 30 14:13:51 crc kubenswrapper[4840]: I0930 14:13:51.099017 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1650244d-222b-4c4e-80f8-720393e7eb67-config-data-custom\") pod \"1650244d-222b-4c4e-80f8-720393e7eb67\" (UID: \"1650244d-222b-4c4e-80f8-720393e7eb67\") " Sep 30 14:13:51 crc kubenswrapper[4840]: I0930 14:13:51.099053 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4z2wj\" (UniqueName: \"kubernetes.io/projected/1650244d-222b-4c4e-80f8-720393e7eb67-kube-api-access-4z2wj\") pod \"1650244d-222b-4c4e-80f8-720393e7eb67\" (UID: \"1650244d-222b-4c4e-80f8-720393e7eb67\") " Sep 30 14:13:51 crc kubenswrapper[4840]: I0930 14:13:51.104001 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1650244d-222b-4c4e-80f8-720393e7eb67-logs" (OuterVolumeSpecName: "logs") pod "1650244d-222b-4c4e-80f8-720393e7eb67" (UID: "1650244d-222b-4c4e-80f8-720393e7eb67"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:13:51 crc kubenswrapper[4840]: I0930 14:13:51.105707 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1650244d-222b-4c4e-80f8-720393e7eb67-kube-api-access-4z2wj" (OuterVolumeSpecName: "kube-api-access-4z2wj") pod "1650244d-222b-4c4e-80f8-720393e7eb67" (UID: "1650244d-222b-4c4e-80f8-720393e7eb67"). InnerVolumeSpecName "kube-api-access-4z2wj". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:13:51 crc kubenswrapper[4840]: I0930 14:13:51.108802 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1650244d-222b-4c4e-80f8-720393e7eb67-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1650244d-222b-4c4e-80f8-720393e7eb67" (UID: "1650244d-222b-4c4e-80f8-720393e7eb67"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:51 crc kubenswrapper[4840]: I0930 14:13:51.128718 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1650244d-222b-4c4e-80f8-720393e7eb67-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1650244d-222b-4c4e-80f8-720393e7eb67" (UID: "1650244d-222b-4c4e-80f8-720393e7eb67"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:51 crc kubenswrapper[4840]: I0930 14:13:51.190226 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1650244d-222b-4c4e-80f8-720393e7eb67-config-data" (OuterVolumeSpecName: "config-data") pod "1650244d-222b-4c4e-80f8-720393e7eb67" (UID: "1650244d-222b-4c4e-80f8-720393e7eb67"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:13:51 crc kubenswrapper[4840]: I0930 14:13:51.201206 4840 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1650244d-222b-4c4e-80f8-720393e7eb67-config-data-custom\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:51 crc kubenswrapper[4840]: I0930 14:13:51.201238 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4z2wj\" (UniqueName: \"kubernetes.io/projected/1650244d-222b-4c4e-80f8-720393e7eb67-kube-api-access-4z2wj\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:51 crc kubenswrapper[4840]: I0930 14:13:51.201254 4840 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1650244d-222b-4c4e-80f8-720393e7eb67-logs\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:51 crc kubenswrapper[4840]: I0930 14:13:51.201265 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1650244d-222b-4c4e-80f8-720393e7eb67-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:51 crc kubenswrapper[4840]: I0930 14:13:51.201276 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1650244d-222b-4c4e-80f8-720393e7eb67-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:13:51 crc kubenswrapper[4840]: I0930 14:13:51.514623 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Sep 30 14:13:51 crc kubenswrapper[4840]: I0930 14:13:51.553045 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-59645bff48-kxnsb" Sep 30 14:13:51 crc kubenswrapper[4840]: I0930 14:13:51.553103 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-59645bff48-kxnsb" event={"ID":"1650244d-222b-4c4e-80f8-720393e7eb67","Type":"ContainerDied","Data":"97be88e5e433c272931edd0cb1c41d8d463cffabc5ac517bbe7ea19f723986a9"} Sep 30 14:13:51 crc kubenswrapper[4840]: I0930 14:13:51.553200 4840 scope.go:117] "RemoveContainer" containerID="9bbeaf4ecaa27d44f6cc27c9d44e044a6e6f58d68591aa17c80309694e88e34e" Sep 30 14:13:51 crc kubenswrapper[4840]: I0930 14:13:51.554600 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"e234012b-317e-4c79-81a3-3991e1296058","Type":"ContainerStarted","Data":"5c8fdaab52edd4dfeda0187bffc9c2521909fbdb1d9028808b8ad3a32f98ca4d"} Sep 30 14:13:51 crc kubenswrapper[4840]: I0930 14:13:51.585393 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-59645bff48-kxnsb"] Sep 30 14:13:51 crc kubenswrapper[4840]: I0930 14:13:51.587769 4840 scope.go:117] "RemoveContainer" containerID="af005a4cb3f5c14bbf6ea4d65b23bee0a91d02cf2ba5b71bf0fd3c936f7904a4" Sep 30 14:13:51 crc kubenswrapper[4840]: I0930 14:13:51.592726 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-59645bff48-kxnsb"] Sep 30 14:13:51 crc kubenswrapper[4840]: I0930 14:13:51.871856 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:13:51 crc kubenswrapper[4840]: I0930 14:13:51.871920 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:13:51 crc kubenswrapper[4840]: I0930 14:13:51.871967 4840 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-747gk" Sep 30 14:13:51 crc kubenswrapper[4840]: I0930 14:13:51.872647 4840 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6060b9705d9c27f06e58cff4fd0a3ee141a9c798fa228092898474533d2401f1"} pod="openshift-machine-config-operator/machine-config-daemon-747gk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 14:13:51 crc kubenswrapper[4840]: I0930 14:13:51.872703 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" containerID="cri-o://6060b9705d9c27f06e58cff4fd0a3ee141a9c798fa228092898474533d2401f1" gracePeriod=600 Sep 30 14:13:52 crc kubenswrapper[4840]: I0930 14:13:52.126223 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1650244d-222b-4c4e-80f8-720393e7eb67" path="/var/lib/kubelet/pods/1650244d-222b-4c4e-80f8-720393e7eb67/volumes" Sep 30 14:13:52 crc kubenswrapper[4840]: I0930 14:13:52.569526 4840 generic.go:334] "Generic (PLEG): container finished" podID="10e8b890-7f20-4a36-8e03-898620cf599a" containerID="6060b9705d9c27f06e58cff4fd0a3ee141a9c798fa228092898474533d2401f1" exitCode=0 Sep 30 14:13:52 crc kubenswrapper[4840]: I0930 14:13:52.569603 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" event={"ID":"10e8b890-7f20-4a36-8e03-898620cf599a","Type":"ContainerDied","Data":"6060b9705d9c27f06e58cff4fd0a3ee141a9c798fa228092898474533d2401f1"} Sep 30 14:13:52 crc kubenswrapper[4840]: I0930 14:13:52.569831 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" event={"ID":"10e8b890-7f20-4a36-8e03-898620cf599a","Type":"ContainerStarted","Data":"5170b28c065e5cc00d1164411d96a8eb01e1860be6b5701def2ab76e6fb63860"} Sep 30 14:13:52 crc kubenswrapper[4840]: I0930 14:13:52.569853 4840 scope.go:117] "RemoveContainer" containerID="999ab6436a76ee1df695f7a3673950dda45d0789af244839074f28ed47e2534b" Sep 30 14:13:52 crc kubenswrapper[4840]: I0930 14:13:52.878287 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.084699 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.085473 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9338f796-fcc1-4b71-8c3e-d48d5234d83d" containerName="ceilometer-central-agent" containerID="cri-o://c0b92bbe8cf99b3b221a23959cc07d7ec9173afe14a9f6f6f88255772435471a" gracePeriod=30 Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.088515 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9338f796-fcc1-4b71-8c3e-d48d5234d83d" containerName="sg-core" containerID="cri-o://834fd6078b398ee02a6379918df8e22ceb2b913854de1e041615c348cdacb890" gracePeriod=30 Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.088598 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9338f796-fcc1-4b71-8c3e-d48d5234d83d" containerName="ceilometer-notification-agent" containerID="cri-o://9aa6a9fbf32866a7f01e8a877ea38fa3904444dda11461efa18674c9d1785f61" gracePeriod=30 Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.088655 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9338f796-fcc1-4b71-8c3e-d48d5234d83d" containerName="proxy-httpd" containerID="cri-o://d234dbb728a108e0737708d70abef8344965f9b6ad50d4052e6ac805d36e26da" gracePeriod=30 Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.102160 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="9338f796-fcc1-4b71-8c3e-d48d5234d83d" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.167:3000/\": EOF" Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.449269 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-5d586fc9-9z92d"] Sep 30 14:13:56 crc kubenswrapper[4840]: E0930 14:13:56.450041 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1650244d-222b-4c4e-80f8-720393e7eb67" containerName="barbican-api" Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.450064 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="1650244d-222b-4c4e-80f8-720393e7eb67" containerName="barbican-api" Sep 30 14:13:56 crc kubenswrapper[4840]: E0930 14:13:56.450083 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1650244d-222b-4c4e-80f8-720393e7eb67" containerName="barbican-api-log" Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.450091 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="1650244d-222b-4c4e-80f8-720393e7eb67" containerName="barbican-api-log" Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.450286 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="1650244d-222b-4c4e-80f8-720393e7eb67" containerName="barbican-api-log" Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.450322 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="1650244d-222b-4c4e-80f8-720393e7eb67" containerName="barbican-api" Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.451592 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5d586fc9-9z92d" Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.453100 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.453857 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.465917 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.467728 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5d586fc9-9z92d"] Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.602669 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/84813c70-f94c-4556-a498-10e17fb354f1-etc-swift\") pod \"swift-proxy-5d586fc9-9z92d\" (UID: \"84813c70-f94c-4556-a498-10e17fb354f1\") " pod="openstack/swift-proxy-5d586fc9-9z92d" Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.602920 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84813c70-f94c-4556-a498-10e17fb354f1-run-httpd\") pod \"swift-proxy-5d586fc9-9z92d\" (UID: \"84813c70-f94c-4556-a498-10e17fb354f1\") " pod="openstack/swift-proxy-5d586fc9-9z92d" Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.603008 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84813c70-f94c-4556-a498-10e17fb354f1-combined-ca-bundle\") pod \"swift-proxy-5d586fc9-9z92d\" (UID: \"84813c70-f94c-4556-a498-10e17fb354f1\") " pod="openstack/swift-proxy-5d586fc9-9z92d" Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.603120 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/84813c70-f94c-4556-a498-10e17fb354f1-internal-tls-certs\") pod \"swift-proxy-5d586fc9-9z92d\" (UID: \"84813c70-f94c-4556-a498-10e17fb354f1\") " pod="openstack/swift-proxy-5d586fc9-9z92d" Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.603328 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84813c70-f94c-4556-a498-10e17fb354f1-log-httpd\") pod \"swift-proxy-5d586fc9-9z92d\" (UID: \"84813c70-f94c-4556-a498-10e17fb354f1\") " pod="openstack/swift-proxy-5d586fc9-9z92d" Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.603447 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84813c70-f94c-4556-a498-10e17fb354f1-config-data\") pod \"swift-proxy-5d586fc9-9z92d\" (UID: \"84813c70-f94c-4556-a498-10e17fb354f1\") " pod="openstack/swift-proxy-5d586fc9-9z92d" Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.603501 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/84813c70-f94c-4556-a498-10e17fb354f1-public-tls-certs\") pod \"swift-proxy-5d586fc9-9z92d\" (UID: \"84813c70-f94c-4556-a498-10e17fb354f1\") " pod="openstack/swift-proxy-5d586fc9-9z92d" Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.603544 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72zh7\" (UniqueName: \"kubernetes.io/projected/84813c70-f94c-4556-a498-10e17fb354f1-kube-api-access-72zh7\") pod \"swift-proxy-5d586fc9-9z92d\" (UID: \"84813c70-f94c-4556-a498-10e17fb354f1\") " pod="openstack/swift-proxy-5d586fc9-9z92d" Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.620508 4840 generic.go:334] "Generic (PLEG): container finished" podID="9338f796-fcc1-4b71-8c3e-d48d5234d83d" containerID="d234dbb728a108e0737708d70abef8344965f9b6ad50d4052e6ac805d36e26da" exitCode=0 Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.620537 4840 generic.go:334] "Generic (PLEG): container finished" podID="9338f796-fcc1-4b71-8c3e-d48d5234d83d" containerID="834fd6078b398ee02a6379918df8e22ceb2b913854de1e041615c348cdacb890" exitCode=2 Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.620544 4840 generic.go:334] "Generic (PLEG): container finished" podID="9338f796-fcc1-4b71-8c3e-d48d5234d83d" containerID="c0b92bbe8cf99b3b221a23959cc07d7ec9173afe14a9f6f6f88255772435471a" exitCode=0 Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.620583 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9338f796-fcc1-4b71-8c3e-d48d5234d83d","Type":"ContainerDied","Data":"d234dbb728a108e0737708d70abef8344965f9b6ad50d4052e6ac805d36e26da"} Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.620626 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9338f796-fcc1-4b71-8c3e-d48d5234d83d","Type":"ContainerDied","Data":"834fd6078b398ee02a6379918df8e22ceb2b913854de1e041615c348cdacb890"} Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.620637 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9338f796-fcc1-4b71-8c3e-d48d5234d83d","Type":"ContainerDied","Data":"c0b92bbe8cf99b3b221a23959cc07d7ec9173afe14a9f6f6f88255772435471a"} Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.704619 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84813c70-f94c-4556-a498-10e17fb354f1-combined-ca-bundle\") pod \"swift-proxy-5d586fc9-9z92d\" (UID: \"84813c70-f94c-4556-a498-10e17fb354f1\") " pod="openstack/swift-proxy-5d586fc9-9z92d" Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.704688 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/84813c70-f94c-4556-a498-10e17fb354f1-internal-tls-certs\") pod \"swift-proxy-5d586fc9-9z92d\" (UID: \"84813c70-f94c-4556-a498-10e17fb354f1\") " pod="openstack/swift-proxy-5d586fc9-9z92d" Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.704718 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84813c70-f94c-4556-a498-10e17fb354f1-log-httpd\") pod \"swift-proxy-5d586fc9-9z92d\" (UID: \"84813c70-f94c-4556-a498-10e17fb354f1\") " pod="openstack/swift-proxy-5d586fc9-9z92d" Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.704743 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84813c70-f94c-4556-a498-10e17fb354f1-config-data\") pod \"swift-proxy-5d586fc9-9z92d\" (UID: \"84813c70-f94c-4556-a498-10e17fb354f1\") " pod="openstack/swift-proxy-5d586fc9-9z92d" Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.704763 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/84813c70-f94c-4556-a498-10e17fb354f1-public-tls-certs\") pod \"swift-proxy-5d586fc9-9z92d\" (UID: \"84813c70-f94c-4556-a498-10e17fb354f1\") " pod="openstack/swift-proxy-5d586fc9-9z92d" Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.704784 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72zh7\" (UniqueName: \"kubernetes.io/projected/84813c70-f94c-4556-a498-10e17fb354f1-kube-api-access-72zh7\") pod \"swift-proxy-5d586fc9-9z92d\" (UID: \"84813c70-f94c-4556-a498-10e17fb354f1\") " pod="openstack/swift-proxy-5d586fc9-9z92d" Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.704805 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/84813c70-f94c-4556-a498-10e17fb354f1-etc-swift\") pod \"swift-proxy-5d586fc9-9z92d\" (UID: \"84813c70-f94c-4556-a498-10e17fb354f1\") " pod="openstack/swift-proxy-5d586fc9-9z92d" Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.704870 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84813c70-f94c-4556-a498-10e17fb354f1-run-httpd\") pod \"swift-proxy-5d586fc9-9z92d\" (UID: \"84813c70-f94c-4556-a498-10e17fb354f1\") " pod="openstack/swift-proxy-5d586fc9-9z92d" Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.705351 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84813c70-f94c-4556-a498-10e17fb354f1-run-httpd\") pod \"swift-proxy-5d586fc9-9z92d\" (UID: \"84813c70-f94c-4556-a498-10e17fb354f1\") " pod="openstack/swift-proxy-5d586fc9-9z92d" Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.705695 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84813c70-f94c-4556-a498-10e17fb354f1-log-httpd\") pod \"swift-proxy-5d586fc9-9z92d\" (UID: \"84813c70-f94c-4556-a498-10e17fb354f1\") " pod="openstack/swift-proxy-5d586fc9-9z92d" Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.715299 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84813c70-f94c-4556-a498-10e17fb354f1-combined-ca-bundle\") pod \"swift-proxy-5d586fc9-9z92d\" (UID: \"84813c70-f94c-4556-a498-10e17fb354f1\") " pod="openstack/swift-proxy-5d586fc9-9z92d" Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.716915 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/84813c70-f94c-4556-a498-10e17fb354f1-public-tls-certs\") pod \"swift-proxy-5d586fc9-9z92d\" (UID: \"84813c70-f94c-4556-a498-10e17fb354f1\") " pod="openstack/swift-proxy-5d586fc9-9z92d" Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.718229 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/84813c70-f94c-4556-a498-10e17fb354f1-etc-swift\") pod \"swift-proxy-5d586fc9-9z92d\" (UID: \"84813c70-f94c-4556-a498-10e17fb354f1\") " pod="openstack/swift-proxy-5d586fc9-9z92d" Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.718419 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84813c70-f94c-4556-a498-10e17fb354f1-config-data\") pod \"swift-proxy-5d586fc9-9z92d\" (UID: \"84813c70-f94c-4556-a498-10e17fb354f1\") " pod="openstack/swift-proxy-5d586fc9-9z92d" Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.721790 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/84813c70-f94c-4556-a498-10e17fb354f1-internal-tls-certs\") pod \"swift-proxy-5d586fc9-9z92d\" (UID: \"84813c70-f94c-4556-a498-10e17fb354f1\") " pod="openstack/swift-proxy-5d586fc9-9z92d" Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.725208 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72zh7\" (UniqueName: \"kubernetes.io/projected/84813c70-f94c-4556-a498-10e17fb354f1-kube-api-access-72zh7\") pod \"swift-proxy-5d586fc9-9z92d\" (UID: \"84813c70-f94c-4556-a498-10e17fb354f1\") " pod="openstack/swift-proxy-5d586fc9-9z92d" Sep 30 14:13:56 crc kubenswrapper[4840]: I0930 14:13:56.768692 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5d586fc9-9z92d" Sep 30 14:13:57 crc kubenswrapper[4840]: I0930 14:13:57.432905 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-xgj9n"] Sep 30 14:13:57 crc kubenswrapper[4840]: I0930 14:13:57.434621 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-xgj9n" Sep 30 14:13:57 crc kubenswrapper[4840]: I0930 14:13:57.442823 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-xgj9n"] Sep 30 14:13:57 crc kubenswrapper[4840]: I0930 14:13:57.521106 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfj44\" (UniqueName: \"kubernetes.io/projected/e00317b4-982d-45c8-b5d6-6b69e4d1d0a7-kube-api-access-nfj44\") pod \"nova-api-db-create-xgj9n\" (UID: \"e00317b4-982d-45c8-b5d6-6b69e4d1d0a7\") " pod="openstack/nova-api-db-create-xgj9n" Sep 30 14:13:57 crc kubenswrapper[4840]: I0930 14:13:57.529700 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-r8kd7"] Sep 30 14:13:57 crc kubenswrapper[4840]: I0930 14:13:57.535649 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-r8kd7" Sep 30 14:13:57 crc kubenswrapper[4840]: I0930 14:13:57.554092 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-r8kd7"] Sep 30 14:13:57 crc kubenswrapper[4840]: I0930 14:13:57.632749 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgkgh\" (UniqueName: \"kubernetes.io/projected/12f02cd8-eb57-4557-b08a-1fc38f1f49fd-kube-api-access-pgkgh\") pod \"nova-cell0-db-create-r8kd7\" (UID: \"12f02cd8-eb57-4557-b08a-1fc38f1f49fd\") " pod="openstack/nova-cell0-db-create-r8kd7" Sep 30 14:13:57 crc kubenswrapper[4840]: I0930 14:13:57.635897 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfj44\" (UniqueName: \"kubernetes.io/projected/e00317b4-982d-45c8-b5d6-6b69e4d1d0a7-kube-api-access-nfj44\") pod \"nova-api-db-create-xgj9n\" (UID: \"e00317b4-982d-45c8-b5d6-6b69e4d1d0a7\") " pod="openstack/nova-api-db-create-xgj9n" Sep 30 14:13:57 crc kubenswrapper[4840]: I0930 14:13:57.642107 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-g7d5k"] Sep 30 14:13:57 crc kubenswrapper[4840]: I0930 14:13:57.647181 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-g7d5k" Sep 30 14:13:57 crc kubenswrapper[4840]: I0930 14:13:57.652071 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-g7d5k"] Sep 30 14:13:57 crc kubenswrapper[4840]: I0930 14:13:57.690041 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfj44\" (UniqueName: \"kubernetes.io/projected/e00317b4-982d-45c8-b5d6-6b69e4d1d0a7-kube-api-access-nfj44\") pod \"nova-api-db-create-xgj9n\" (UID: \"e00317b4-982d-45c8-b5d6-6b69e4d1d0a7\") " pod="openstack/nova-api-db-create-xgj9n" Sep 30 14:13:57 crc kubenswrapper[4840]: I0930 14:13:57.737952 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4jqn\" (UniqueName: \"kubernetes.io/projected/252dac07-3c44-4264-a488-cca59b98e516-kube-api-access-r4jqn\") pod \"nova-cell1-db-create-g7d5k\" (UID: \"252dac07-3c44-4264-a488-cca59b98e516\") " pod="openstack/nova-cell1-db-create-g7d5k" Sep 30 14:13:57 crc kubenswrapper[4840]: I0930 14:13:57.738080 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgkgh\" (UniqueName: \"kubernetes.io/projected/12f02cd8-eb57-4557-b08a-1fc38f1f49fd-kube-api-access-pgkgh\") pod \"nova-cell0-db-create-r8kd7\" (UID: \"12f02cd8-eb57-4557-b08a-1fc38f1f49fd\") " pod="openstack/nova-cell0-db-create-r8kd7" Sep 30 14:13:57 crc kubenswrapper[4840]: I0930 14:13:57.750963 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-xgj9n" Sep 30 14:13:57 crc kubenswrapper[4840]: I0930 14:13:57.756747 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgkgh\" (UniqueName: \"kubernetes.io/projected/12f02cd8-eb57-4557-b08a-1fc38f1f49fd-kube-api-access-pgkgh\") pod \"nova-cell0-db-create-r8kd7\" (UID: \"12f02cd8-eb57-4557-b08a-1fc38f1f49fd\") " pod="openstack/nova-cell0-db-create-r8kd7" Sep 30 14:13:57 crc kubenswrapper[4840]: I0930 14:13:57.839959 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4jqn\" (UniqueName: \"kubernetes.io/projected/252dac07-3c44-4264-a488-cca59b98e516-kube-api-access-r4jqn\") pod \"nova-cell1-db-create-g7d5k\" (UID: \"252dac07-3c44-4264-a488-cca59b98e516\") " pod="openstack/nova-cell1-db-create-g7d5k" Sep 30 14:13:57 crc kubenswrapper[4840]: I0930 14:13:57.856412 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4jqn\" (UniqueName: \"kubernetes.io/projected/252dac07-3c44-4264-a488-cca59b98e516-kube-api-access-r4jqn\") pod \"nova-cell1-db-create-g7d5k\" (UID: \"252dac07-3c44-4264-a488-cca59b98e516\") " pod="openstack/nova-cell1-db-create-g7d5k" Sep 30 14:13:57 crc kubenswrapper[4840]: I0930 14:13:57.895359 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-r8kd7" Sep 30 14:13:58 crc kubenswrapper[4840]: I0930 14:13:58.023690 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-g7d5k" Sep 30 14:13:58 crc kubenswrapper[4840]: I0930 14:13:58.092845 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Sep 30 14:13:58 crc kubenswrapper[4840]: I0930 14:13:58.594830 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5595d79c68-dgsth" podUID="5e9b87b2-4836-4857-8514-b2dcfe4de1f5" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Sep 30 14:13:59 crc kubenswrapper[4840]: I0930 14:13:59.281131 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6cbd5fdcfb-hn4pn" Sep 30 14:13:59 crc kubenswrapper[4840]: I0930 14:13:59.354509 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6cbd5fdcfb-hn4pn" Sep 30 14:14:00 crc kubenswrapper[4840]: I0930 14:14:00.688980 4840 generic.go:334] "Generic (PLEG): container finished" podID="9338f796-fcc1-4b71-8c3e-d48d5234d83d" containerID="9aa6a9fbf32866a7f01e8a877ea38fa3904444dda11461efa18674c9d1785f61" exitCode=0 Sep 30 14:14:00 crc kubenswrapper[4840]: I0930 14:14:00.689064 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9338f796-fcc1-4b71-8c3e-d48d5234d83d","Type":"ContainerDied","Data":"9aa6a9fbf32866a7f01e8a877ea38fa3904444dda11461efa18674c9d1785f61"} Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.328203 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.407093 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9338f796-fcc1-4b71-8c3e-d48d5234d83d-log-httpd\") pod \"9338f796-fcc1-4b71-8c3e-d48d5234d83d\" (UID: \"9338f796-fcc1-4b71-8c3e-d48d5234d83d\") " Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.407201 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9338f796-fcc1-4b71-8c3e-d48d5234d83d-run-httpd\") pod \"9338f796-fcc1-4b71-8c3e-d48d5234d83d\" (UID: \"9338f796-fcc1-4b71-8c3e-d48d5234d83d\") " Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.407287 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9338f796-fcc1-4b71-8c3e-d48d5234d83d-combined-ca-bundle\") pod \"9338f796-fcc1-4b71-8c3e-d48d5234d83d\" (UID: \"9338f796-fcc1-4b71-8c3e-d48d5234d83d\") " Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.407324 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9338f796-fcc1-4b71-8c3e-d48d5234d83d-config-data\") pod \"9338f796-fcc1-4b71-8c3e-d48d5234d83d\" (UID: \"9338f796-fcc1-4b71-8c3e-d48d5234d83d\") " Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.407353 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9338f796-fcc1-4b71-8c3e-d48d5234d83d-scripts\") pod \"9338f796-fcc1-4b71-8c3e-d48d5234d83d\" (UID: \"9338f796-fcc1-4b71-8c3e-d48d5234d83d\") " Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.407396 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdqzz\" (UniqueName: \"kubernetes.io/projected/9338f796-fcc1-4b71-8c3e-d48d5234d83d-kube-api-access-fdqzz\") pod \"9338f796-fcc1-4b71-8c3e-d48d5234d83d\" (UID: \"9338f796-fcc1-4b71-8c3e-d48d5234d83d\") " Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.407461 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9338f796-fcc1-4b71-8c3e-d48d5234d83d-sg-core-conf-yaml\") pod \"9338f796-fcc1-4b71-8c3e-d48d5234d83d\" (UID: \"9338f796-fcc1-4b71-8c3e-d48d5234d83d\") " Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.408625 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9338f796-fcc1-4b71-8c3e-d48d5234d83d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9338f796-fcc1-4b71-8c3e-d48d5234d83d" (UID: "9338f796-fcc1-4b71-8c3e-d48d5234d83d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.408782 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9338f796-fcc1-4b71-8c3e-d48d5234d83d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9338f796-fcc1-4b71-8c3e-d48d5234d83d" (UID: "9338f796-fcc1-4b71-8c3e-d48d5234d83d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.412308 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9338f796-fcc1-4b71-8c3e-d48d5234d83d-kube-api-access-fdqzz" (OuterVolumeSpecName: "kube-api-access-fdqzz") pod "9338f796-fcc1-4b71-8c3e-d48d5234d83d" (UID: "9338f796-fcc1-4b71-8c3e-d48d5234d83d"). InnerVolumeSpecName "kube-api-access-fdqzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.412516 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9338f796-fcc1-4b71-8c3e-d48d5234d83d-scripts" (OuterVolumeSpecName: "scripts") pod "9338f796-fcc1-4b71-8c3e-d48d5234d83d" (UID: "9338f796-fcc1-4b71-8c3e-d48d5234d83d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.449228 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9338f796-fcc1-4b71-8c3e-d48d5234d83d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9338f796-fcc1-4b71-8c3e-d48d5234d83d" (UID: "9338f796-fcc1-4b71-8c3e-d48d5234d83d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.501482 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9338f796-fcc1-4b71-8c3e-d48d5234d83d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9338f796-fcc1-4b71-8c3e-d48d5234d83d" (UID: "9338f796-fcc1-4b71-8c3e-d48d5234d83d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.509752 4840 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9338f796-fcc1-4b71-8c3e-d48d5234d83d-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.509787 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9338f796-fcc1-4b71-8c3e-d48d5234d83d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.509803 4840 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9338f796-fcc1-4b71-8c3e-d48d5234d83d-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.509814 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdqzz\" (UniqueName: \"kubernetes.io/projected/9338f796-fcc1-4b71-8c3e-d48d5234d83d-kube-api-access-fdqzz\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.509826 4840 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9338f796-fcc1-4b71-8c3e-d48d5234d83d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.509836 4840 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9338f796-fcc1-4b71-8c3e-d48d5234d83d-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.521495 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9338f796-fcc1-4b71-8c3e-d48d5234d83d-config-data" (OuterVolumeSpecName: "config-data") pod "9338f796-fcc1-4b71-8c3e-d48d5234d83d" (UID: "9338f796-fcc1-4b71-8c3e-d48d5234d83d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.611482 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9338f796-fcc1-4b71-8c3e-d48d5234d83d-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.639731 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.639993 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="b96c744c-10e9-4c0c-9ece-069c0c90e67a" containerName="glance-log" containerID="cri-o://f19c53a63f714f8e0a84ed1833fdf7a816a89d7939ff5aad1ade067169e78a35" gracePeriod=30 Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.640190 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="b96c744c-10e9-4c0c-9ece-069c0c90e67a" containerName="glance-httpd" containerID="cri-o://3c95288a14cc91765a03616025aa0e85a31db07cc1cc694550d048ce584c730b" gracePeriod=30 Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.650357 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-r8kd7"] Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.662020 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-xgj9n"] Sep 30 14:14:01 crc kubenswrapper[4840]: W0930 14:14:01.666768 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12f02cd8_eb57_4557_b08a_1fc38f1f49fd.slice/crio-653b193f22f430e8ca99c1128b2657125af254f96079c3da5b0b3f550f30e14e WatchSource:0}: Error finding container 653b193f22f430e8ca99c1128b2657125af254f96079c3da5b0b3f550f30e14e: Status 404 returned error can't find the container with id 653b193f22f430e8ca99c1128b2657125af254f96079c3da5b0b3f550f30e14e Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.709275 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-xgj9n" event={"ID":"e00317b4-982d-45c8-b5d6-6b69e4d1d0a7","Type":"ContainerStarted","Data":"f8724c2a43faba418d07f851986fc1ad91fdce3e01139f7180252d29defa674f"} Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.716852 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9338f796-fcc1-4b71-8c3e-d48d5234d83d","Type":"ContainerDied","Data":"498887b9fc0395006aaf1d420b15b949d3c567b64f943cc13c1ffc1bc991807f"} Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.716905 4840 scope.go:117] "RemoveContainer" containerID="d234dbb728a108e0737708d70abef8344965f9b6ad50d4052e6ac805d36e26da" Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.717029 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.734161 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"e234012b-317e-4c79-81a3-3991e1296058","Type":"ContainerStarted","Data":"934ebe8804c6a0e86e5e23b239ec9637b4b73ba9f57dc5ea7e41118f8da920ef"} Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.739257 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-r8kd7" event={"ID":"12f02cd8-eb57-4557-b08a-1fc38f1f49fd","Type":"ContainerStarted","Data":"653b193f22f430e8ca99c1128b2657125af254f96079c3da5b0b3f550f30e14e"} Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.765652 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.150722657 podStartE2EDuration="11.765627265s" podCreationTimestamp="2025-09-30 14:13:50 +0000 UTC" firstStartedPulling="2025-09-30 14:13:51.532541086 +0000 UTC m=+1060.161627509" lastFinishedPulling="2025-09-30 14:14:01.147445694 +0000 UTC m=+1069.776532117" observedRunningTime="2025-09-30 14:14:01.746893306 +0000 UTC m=+1070.375979719" watchObservedRunningTime="2025-09-30 14:14:01.765627265 +0000 UTC m=+1070.394713698" Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.790703 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5d586fc9-9z92d"] Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.797926 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-g7d5k"] Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.798225 4840 scope.go:117] "RemoveContainer" containerID="834fd6078b398ee02a6379918df8e22ceb2b913854de1e041615c348cdacb890" Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.817282 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.830207 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:14:01 crc kubenswrapper[4840]: W0930 14:14:01.833424 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod252dac07_3c44_4264_a488_cca59b98e516.slice/crio-542c6fc73de8f680a51e698031ab122ff72f4bfeb8a232a897a1576beceec4bf WatchSource:0}: Error finding container 542c6fc73de8f680a51e698031ab122ff72f4bfeb8a232a897a1576beceec4bf: Status 404 returned error can't find the container with id 542c6fc73de8f680a51e698031ab122ff72f4bfeb8a232a897a1576beceec4bf Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.840447 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:14:01 crc kubenswrapper[4840]: E0930 14:14:01.840784 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9338f796-fcc1-4b71-8c3e-d48d5234d83d" containerName="ceilometer-central-agent" Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.840798 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="9338f796-fcc1-4b71-8c3e-d48d5234d83d" containerName="ceilometer-central-agent" Sep 30 14:14:01 crc kubenswrapper[4840]: E0930 14:14:01.840816 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9338f796-fcc1-4b71-8c3e-d48d5234d83d" containerName="ceilometer-notification-agent" Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.840822 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="9338f796-fcc1-4b71-8c3e-d48d5234d83d" containerName="ceilometer-notification-agent" Sep 30 14:14:01 crc kubenswrapper[4840]: E0930 14:14:01.840832 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9338f796-fcc1-4b71-8c3e-d48d5234d83d" containerName="sg-core" Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.840838 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="9338f796-fcc1-4b71-8c3e-d48d5234d83d" containerName="sg-core" Sep 30 14:14:01 crc kubenswrapper[4840]: E0930 14:14:01.840856 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9338f796-fcc1-4b71-8c3e-d48d5234d83d" containerName="proxy-httpd" Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.840862 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="9338f796-fcc1-4b71-8c3e-d48d5234d83d" containerName="proxy-httpd" Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.841073 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="9338f796-fcc1-4b71-8c3e-d48d5234d83d" containerName="sg-core" Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.841094 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="9338f796-fcc1-4b71-8c3e-d48d5234d83d" containerName="ceilometer-notification-agent" Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.841104 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="9338f796-fcc1-4b71-8c3e-d48d5234d83d" containerName="ceilometer-central-agent" Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.841122 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="9338f796-fcc1-4b71-8c3e-d48d5234d83d" containerName="proxy-httpd" Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.844043 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.845056 4840 scope.go:117] "RemoveContainer" containerID="9aa6a9fbf32866a7f01e8a877ea38fa3904444dda11461efa18674c9d1785f61" Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.853403 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.853724 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.854039 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.876774 4840 scope.go:117] "RemoveContainer" containerID="c0b92bbe8cf99b3b221a23959cc07d7ec9173afe14a9f6f6f88255772435471a" Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.925935 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-run-httpd\") pod \"ceilometer-0\" (UID: \"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf\") " pod="openstack/ceilometer-0" Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.926196 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9q7qd\" (UniqueName: \"kubernetes.io/projected/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-kube-api-access-9q7qd\") pod \"ceilometer-0\" (UID: \"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf\") " pod="openstack/ceilometer-0" Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.926299 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf\") " pod="openstack/ceilometer-0" Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.926438 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-scripts\") pod \"ceilometer-0\" (UID: \"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf\") " pod="openstack/ceilometer-0" Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.926703 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf\") " pod="openstack/ceilometer-0" Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.926873 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-config-data\") pod \"ceilometer-0\" (UID: \"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf\") " pod="openstack/ceilometer-0" Sep 30 14:14:01 crc kubenswrapper[4840]: I0930 14:14:01.927085 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-log-httpd\") pod \"ceilometer-0\" (UID: \"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf\") " pod="openstack/ceilometer-0" Sep 30 14:14:02 crc kubenswrapper[4840]: I0930 14:14:02.028928 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-log-httpd\") pod \"ceilometer-0\" (UID: \"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf\") " pod="openstack/ceilometer-0" Sep 30 14:14:02 crc kubenswrapper[4840]: I0930 14:14:02.029025 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-run-httpd\") pod \"ceilometer-0\" (UID: \"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf\") " pod="openstack/ceilometer-0" Sep 30 14:14:02 crc kubenswrapper[4840]: I0930 14:14:02.029044 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9q7qd\" (UniqueName: \"kubernetes.io/projected/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-kube-api-access-9q7qd\") pod \"ceilometer-0\" (UID: \"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf\") " pod="openstack/ceilometer-0" Sep 30 14:14:02 crc kubenswrapper[4840]: I0930 14:14:02.029065 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf\") " pod="openstack/ceilometer-0" Sep 30 14:14:02 crc kubenswrapper[4840]: I0930 14:14:02.029101 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-scripts\") pod \"ceilometer-0\" (UID: \"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf\") " pod="openstack/ceilometer-0" Sep 30 14:14:02 crc kubenswrapper[4840]: I0930 14:14:02.029139 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf\") " pod="openstack/ceilometer-0" Sep 30 14:14:02 crc kubenswrapper[4840]: I0930 14:14:02.029184 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-config-data\") pod \"ceilometer-0\" (UID: \"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf\") " pod="openstack/ceilometer-0" Sep 30 14:14:02 crc kubenswrapper[4840]: I0930 14:14:02.030039 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-log-httpd\") pod \"ceilometer-0\" (UID: \"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf\") " pod="openstack/ceilometer-0" Sep 30 14:14:02 crc kubenswrapper[4840]: I0930 14:14:02.030363 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-run-httpd\") pod \"ceilometer-0\" (UID: \"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf\") " pod="openstack/ceilometer-0" Sep 30 14:14:02 crc kubenswrapper[4840]: I0930 14:14:02.033648 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-scripts\") pod \"ceilometer-0\" (UID: \"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf\") " pod="openstack/ceilometer-0" Sep 30 14:14:02 crc kubenswrapper[4840]: I0930 14:14:02.034203 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-config-data\") pod \"ceilometer-0\" (UID: \"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf\") " pod="openstack/ceilometer-0" Sep 30 14:14:02 crc kubenswrapper[4840]: I0930 14:14:02.034619 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf\") " pod="openstack/ceilometer-0" Sep 30 14:14:02 crc kubenswrapper[4840]: I0930 14:14:02.036221 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf\") " pod="openstack/ceilometer-0" Sep 30 14:14:02 crc kubenswrapper[4840]: I0930 14:14:02.049913 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9q7qd\" (UniqueName: \"kubernetes.io/projected/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-kube-api-access-9q7qd\") pod \"ceilometer-0\" (UID: \"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf\") " pod="openstack/ceilometer-0" Sep 30 14:14:02 crc kubenswrapper[4840]: I0930 14:14:02.127193 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9338f796-fcc1-4b71-8c3e-d48d5234d83d" path="/var/lib/kubelet/pods/9338f796-fcc1-4b71-8c3e-d48d5234d83d/volumes" Sep 30 14:14:02 crc kubenswrapper[4840]: I0930 14:14:02.178384 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:14:02 crc kubenswrapper[4840]: I0930 14:14:02.506072 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:14:02 crc kubenswrapper[4840]: I0930 14:14:02.506588 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="c499ea9a-7047-493b-a529-1b5115119b25" containerName="glance-log" containerID="cri-o://03648621acfcfc360502f4b3af076777d7d5f8a4f30a001515922855a9ab1fe9" gracePeriod=30 Sep 30 14:14:02 crc kubenswrapper[4840]: I0930 14:14:02.507316 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="c499ea9a-7047-493b-a529-1b5115119b25" containerName="glance-httpd" containerID="cri-o://b203574d1cba5c9043bdf32ea9de8523a03e3e11b59553bd30a8431865d24f1d" gracePeriod=30 Sep 30 14:14:02 crc kubenswrapper[4840]: I0930 14:14:02.628891 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:14:02 crc kubenswrapper[4840]: W0930 14:14:02.633812 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4aa3849f_5f45_4c05_ac5a_dc8472d5acaf.slice/crio-d05eee7cc0d5a444a422917ea1dd7c715fb505e754eee48067d533858389f278 WatchSource:0}: Error finding container d05eee7cc0d5a444a422917ea1dd7c715fb505e754eee48067d533858389f278: Status 404 returned error can't find the container with id d05eee7cc0d5a444a422917ea1dd7c715fb505e754eee48067d533858389f278 Sep 30 14:14:02 crc kubenswrapper[4840]: I0930 14:14:02.752691 4840 generic.go:334] "Generic (PLEG): container finished" podID="252dac07-3c44-4264-a488-cca59b98e516" containerID="333296e5f81bf6eb25692aa0848f4ac2b2c64935f8ca8bb9b5804f26e314562f" exitCode=0 Sep 30 14:14:02 crc kubenswrapper[4840]: I0930 14:14:02.752762 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-g7d5k" event={"ID":"252dac07-3c44-4264-a488-cca59b98e516","Type":"ContainerDied","Data":"333296e5f81bf6eb25692aa0848f4ac2b2c64935f8ca8bb9b5804f26e314562f"} Sep 30 14:14:02 crc kubenswrapper[4840]: I0930 14:14:02.752789 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-g7d5k" event={"ID":"252dac07-3c44-4264-a488-cca59b98e516","Type":"ContainerStarted","Data":"542c6fc73de8f680a51e698031ab122ff72f4bfeb8a232a897a1576beceec4bf"} Sep 30 14:14:02 crc kubenswrapper[4840]: I0930 14:14:02.754998 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5d586fc9-9z92d" event={"ID":"84813c70-f94c-4556-a498-10e17fb354f1","Type":"ContainerStarted","Data":"fdbb1a254d40d5e7da956d37f445446791d6360a087dd9bb1a9f79fd85e4dfe8"} Sep 30 14:14:02 crc kubenswrapper[4840]: I0930 14:14:02.755030 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5d586fc9-9z92d" event={"ID":"84813c70-f94c-4556-a498-10e17fb354f1","Type":"ContainerStarted","Data":"c21b9f5b67f6d6fe5d4859ed8f2192ca0f5e29b2b01339d6b066e0c0e2aec8f2"} Sep 30 14:14:02 crc kubenswrapper[4840]: I0930 14:14:02.755040 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5d586fc9-9z92d" event={"ID":"84813c70-f94c-4556-a498-10e17fb354f1","Type":"ContainerStarted","Data":"32f96db1d38884ce7ad623d16268c2cd3e52413c1cbb038e45409c5297b29212"} Sep 30 14:14:02 crc kubenswrapper[4840]: I0930 14:14:02.755136 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5d586fc9-9z92d" Sep 30 14:14:02 crc kubenswrapper[4840]: I0930 14:14:02.755152 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5d586fc9-9z92d" Sep 30 14:14:02 crc kubenswrapper[4840]: I0930 14:14:02.759400 4840 generic.go:334] "Generic (PLEG): container finished" podID="c499ea9a-7047-493b-a529-1b5115119b25" containerID="03648621acfcfc360502f4b3af076777d7d5f8a4f30a001515922855a9ab1fe9" exitCode=143 Sep 30 14:14:02 crc kubenswrapper[4840]: I0930 14:14:02.759490 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c499ea9a-7047-493b-a529-1b5115119b25","Type":"ContainerDied","Data":"03648621acfcfc360502f4b3af076777d7d5f8a4f30a001515922855a9ab1fe9"} Sep 30 14:14:02 crc kubenswrapper[4840]: I0930 14:14:02.763928 4840 generic.go:334] "Generic (PLEG): container finished" podID="12f02cd8-eb57-4557-b08a-1fc38f1f49fd" containerID="d8633210762a0973646719f652972b15063961d632b50a670d911d219e79321e" exitCode=0 Sep 30 14:14:02 crc kubenswrapper[4840]: I0930 14:14:02.764197 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-r8kd7" event={"ID":"12f02cd8-eb57-4557-b08a-1fc38f1f49fd","Type":"ContainerDied","Data":"d8633210762a0973646719f652972b15063961d632b50a670d911d219e79321e"} Sep 30 14:14:02 crc kubenswrapper[4840]: I0930 14:14:02.766673 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf","Type":"ContainerStarted","Data":"d05eee7cc0d5a444a422917ea1dd7c715fb505e754eee48067d533858389f278"} Sep 30 14:14:02 crc kubenswrapper[4840]: I0930 14:14:02.768608 4840 generic.go:334] "Generic (PLEG): container finished" podID="e00317b4-982d-45c8-b5d6-6b69e4d1d0a7" containerID="c1bf996a56aa4ade8ada32f7bee83c4df3bacd4b72b8b401a2c1597ee6724fa5" exitCode=0 Sep 30 14:14:02 crc kubenswrapper[4840]: I0930 14:14:02.768689 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-xgj9n" event={"ID":"e00317b4-982d-45c8-b5d6-6b69e4d1d0a7","Type":"ContainerDied","Data":"c1bf996a56aa4ade8ada32f7bee83c4df3bacd4b72b8b401a2c1597ee6724fa5"} Sep 30 14:14:02 crc kubenswrapper[4840]: I0930 14:14:02.771986 4840 generic.go:334] "Generic (PLEG): container finished" podID="b96c744c-10e9-4c0c-9ece-069c0c90e67a" containerID="f19c53a63f714f8e0a84ed1833fdf7a816a89d7939ff5aad1ade067169e78a35" exitCode=143 Sep 30 14:14:02 crc kubenswrapper[4840]: I0930 14:14:02.772254 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b96c744c-10e9-4c0c-9ece-069c0c90e67a","Type":"ContainerDied","Data":"f19c53a63f714f8e0a84ed1833fdf7a816a89d7939ff5aad1ade067169e78a35"} Sep 30 14:14:02 crc kubenswrapper[4840]: I0930 14:14:02.806694 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-5d586fc9-9z92d" podStartSLOduration=6.806622352 podStartE2EDuration="6.806622352s" podCreationTimestamp="2025-09-30 14:13:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:14:02.78430774 +0000 UTC m=+1071.413394193" watchObservedRunningTime="2025-09-30 14:14:02.806622352 +0000 UTC m=+1071.435708765" Sep 30 14:14:03 crc kubenswrapper[4840]: I0930 14:14:03.689739 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:14:03 crc kubenswrapper[4840]: I0930 14:14:03.782183 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf","Type":"ContainerStarted","Data":"cfcaf10f26382eada414852d9cab5083c54f9e5f2466f123779ce16711278cf9"} Sep 30 14:14:04 crc kubenswrapper[4840]: I0930 14:14:04.395182 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-xgj9n" Sep 30 14:14:04 crc kubenswrapper[4840]: I0930 14:14:04.402029 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-r8kd7" Sep 30 14:14:04 crc kubenswrapper[4840]: I0930 14:14:04.409939 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-g7d5k" Sep 30 14:14:04 crc kubenswrapper[4840]: I0930 14:14:04.480381 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4jqn\" (UniqueName: \"kubernetes.io/projected/252dac07-3c44-4264-a488-cca59b98e516-kube-api-access-r4jqn\") pod \"252dac07-3c44-4264-a488-cca59b98e516\" (UID: \"252dac07-3c44-4264-a488-cca59b98e516\") " Sep 30 14:14:04 crc kubenswrapper[4840]: I0930 14:14:04.480661 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgkgh\" (UniqueName: \"kubernetes.io/projected/12f02cd8-eb57-4557-b08a-1fc38f1f49fd-kube-api-access-pgkgh\") pod \"12f02cd8-eb57-4557-b08a-1fc38f1f49fd\" (UID: \"12f02cd8-eb57-4557-b08a-1fc38f1f49fd\") " Sep 30 14:14:04 crc kubenswrapper[4840]: I0930 14:14:04.480846 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nfj44\" (UniqueName: \"kubernetes.io/projected/e00317b4-982d-45c8-b5d6-6b69e4d1d0a7-kube-api-access-nfj44\") pod \"e00317b4-982d-45c8-b5d6-6b69e4d1d0a7\" (UID: \"e00317b4-982d-45c8-b5d6-6b69e4d1d0a7\") " Sep 30 14:14:04 crc kubenswrapper[4840]: I0930 14:14:04.494735 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12f02cd8-eb57-4557-b08a-1fc38f1f49fd-kube-api-access-pgkgh" (OuterVolumeSpecName: "kube-api-access-pgkgh") pod "12f02cd8-eb57-4557-b08a-1fc38f1f49fd" (UID: "12f02cd8-eb57-4557-b08a-1fc38f1f49fd"). InnerVolumeSpecName "kube-api-access-pgkgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:14:04 crc kubenswrapper[4840]: I0930 14:14:04.499038 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/252dac07-3c44-4264-a488-cca59b98e516-kube-api-access-r4jqn" (OuterVolumeSpecName: "kube-api-access-r4jqn") pod "252dac07-3c44-4264-a488-cca59b98e516" (UID: "252dac07-3c44-4264-a488-cca59b98e516"). InnerVolumeSpecName "kube-api-access-r4jqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:14:04 crc kubenswrapper[4840]: I0930 14:14:04.501190 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e00317b4-982d-45c8-b5d6-6b69e4d1d0a7-kube-api-access-nfj44" (OuterVolumeSpecName: "kube-api-access-nfj44") pod "e00317b4-982d-45c8-b5d6-6b69e4d1d0a7" (UID: "e00317b4-982d-45c8-b5d6-6b69e4d1d0a7"). InnerVolumeSpecName "kube-api-access-nfj44". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:14:04 crc kubenswrapper[4840]: I0930 14:14:04.582939 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4jqn\" (UniqueName: \"kubernetes.io/projected/252dac07-3c44-4264-a488-cca59b98e516-kube-api-access-r4jqn\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:04 crc kubenswrapper[4840]: I0930 14:14:04.582981 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgkgh\" (UniqueName: \"kubernetes.io/projected/12f02cd8-eb57-4557-b08a-1fc38f1f49fd-kube-api-access-pgkgh\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:04 crc kubenswrapper[4840]: I0930 14:14:04.582991 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nfj44\" (UniqueName: \"kubernetes.io/projected/e00317b4-982d-45c8-b5d6-6b69e4d1d0a7-kube-api-access-nfj44\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:04 crc kubenswrapper[4840]: I0930 14:14:04.793457 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-r8kd7" Sep 30 14:14:04 crc kubenswrapper[4840]: I0930 14:14:04.793455 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-r8kd7" event={"ID":"12f02cd8-eb57-4557-b08a-1fc38f1f49fd","Type":"ContainerDied","Data":"653b193f22f430e8ca99c1128b2657125af254f96079c3da5b0b3f550f30e14e"} Sep 30 14:14:04 crc kubenswrapper[4840]: I0930 14:14:04.793603 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="653b193f22f430e8ca99c1128b2657125af254f96079c3da5b0b3f550f30e14e" Sep 30 14:14:04 crc kubenswrapper[4840]: I0930 14:14:04.795434 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-xgj9n" Sep 30 14:14:04 crc kubenswrapper[4840]: I0930 14:14:04.795461 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-xgj9n" event={"ID":"e00317b4-982d-45c8-b5d6-6b69e4d1d0a7","Type":"ContainerDied","Data":"f8724c2a43faba418d07f851986fc1ad91fdce3e01139f7180252d29defa674f"} Sep 30 14:14:04 crc kubenswrapper[4840]: I0930 14:14:04.795485 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f8724c2a43faba418d07f851986fc1ad91fdce3e01139f7180252d29defa674f" Sep 30 14:14:04 crc kubenswrapper[4840]: I0930 14:14:04.797083 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-g7d5k" event={"ID":"252dac07-3c44-4264-a488-cca59b98e516","Type":"ContainerDied","Data":"542c6fc73de8f680a51e698031ab122ff72f4bfeb8a232a897a1576beceec4bf"} Sep 30 14:14:04 crc kubenswrapper[4840]: I0930 14:14:04.797103 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="542c6fc73de8f680a51e698031ab122ff72f4bfeb8a232a897a1576beceec4bf" Sep 30 14:14:04 crc kubenswrapper[4840]: I0930 14:14:04.797144 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-g7d5k" Sep 30 14:14:05 crc kubenswrapper[4840]: I0930 14:14:05.807297 4840 generic.go:334] "Generic (PLEG): container finished" podID="b96c744c-10e9-4c0c-9ece-069c0c90e67a" containerID="3c95288a14cc91765a03616025aa0e85a31db07cc1cc694550d048ce584c730b" exitCode=0 Sep 30 14:14:05 crc kubenswrapper[4840]: I0930 14:14:05.807402 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b96c744c-10e9-4c0c-9ece-069c0c90e67a","Type":"ContainerDied","Data":"3c95288a14cc91765a03616025aa0e85a31db07cc1cc694550d048ce584c730b"} Sep 30 14:14:05 crc kubenswrapper[4840]: I0930 14:14:05.809356 4840 generic.go:334] "Generic (PLEG): container finished" podID="c499ea9a-7047-493b-a529-1b5115119b25" containerID="b203574d1cba5c9043bdf32ea9de8523a03e3e11b59553bd30a8431865d24f1d" exitCode=0 Sep 30 14:14:05 crc kubenswrapper[4840]: I0930 14:14:05.809383 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c499ea9a-7047-493b-a529-1b5115119b25","Type":"ContainerDied","Data":"b203574d1cba5c9043bdf32ea9de8523a03e3e11b59553bd30a8431865d24f1d"} Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.194623 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.311304 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b96c744c-10e9-4c0c-9ece-069c0c90e67a-config-data\") pod \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\" (UID: \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\") " Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.311373 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b96c744c-10e9-4c0c-9ece-069c0c90e67a-scripts\") pod \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\" (UID: \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\") " Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.312054 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b96c744c-10e9-4c0c-9ece-069c0c90e67a-combined-ca-bundle\") pod \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\" (UID: \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\") " Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.312097 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\" (UID: \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\") " Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.312127 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72l8s\" (UniqueName: \"kubernetes.io/projected/b96c744c-10e9-4c0c-9ece-069c0c90e67a-kube-api-access-72l8s\") pod \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\" (UID: \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\") " Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.312172 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b96c744c-10e9-4c0c-9ece-069c0c90e67a-logs\") pod \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\" (UID: \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\") " Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.312189 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b96c744c-10e9-4c0c-9ece-069c0c90e67a-public-tls-certs\") pod \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\" (UID: \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\") " Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.312208 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b96c744c-10e9-4c0c-9ece-069c0c90e67a-httpd-run\") pod \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\" (UID: \"b96c744c-10e9-4c0c-9ece-069c0c90e67a\") " Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.312834 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b96c744c-10e9-4c0c-9ece-069c0c90e67a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b96c744c-10e9-4c0c-9ece-069c0c90e67a" (UID: "b96c744c-10e9-4c0c-9ece-069c0c90e67a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.312962 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b96c744c-10e9-4c0c-9ece-069c0c90e67a-logs" (OuterVolumeSpecName: "logs") pod "b96c744c-10e9-4c0c-9ece-069c0c90e67a" (UID: "b96c744c-10e9-4c0c-9ece-069c0c90e67a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.317130 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "b96c744c-10e9-4c0c-9ece-069c0c90e67a" (UID: "b96c744c-10e9-4c0c-9ece-069c0c90e67a"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.317901 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b96c744c-10e9-4c0c-9ece-069c0c90e67a-kube-api-access-72l8s" (OuterVolumeSpecName: "kube-api-access-72l8s") pod "b96c744c-10e9-4c0c-9ece-069c0c90e67a" (UID: "b96c744c-10e9-4c0c-9ece-069c0c90e67a"). InnerVolumeSpecName "kube-api-access-72l8s". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.324948 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b96c744c-10e9-4c0c-9ece-069c0c90e67a-scripts" (OuterVolumeSpecName: "scripts") pod "b96c744c-10e9-4c0c-9ece-069c0c90e67a" (UID: "b96c744c-10e9-4c0c-9ece-069c0c90e67a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.341663 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b96c744c-10e9-4c0c-9ece-069c0c90e67a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b96c744c-10e9-4c0c-9ece-069c0c90e67a" (UID: "b96c744c-10e9-4c0c-9ece-069c0c90e67a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.364834 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b96c744c-10e9-4c0c-9ece-069c0c90e67a-config-data" (OuterVolumeSpecName: "config-data") pod "b96c744c-10e9-4c0c-9ece-069c0c90e67a" (UID: "b96c744c-10e9-4c0c-9ece-069c0c90e67a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.367293 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b96c744c-10e9-4c0c-9ece-069c0c90e67a-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "b96c744c-10e9-4c0c-9ece-069c0c90e67a" (UID: "b96c744c-10e9-4c0c-9ece-069c0c90e67a"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.414130 4840 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b96c744c-10e9-4c0c-9ece-069c0c90e67a-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.414307 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b96c744c-10e9-4c0c-9ece-069c0c90e67a-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.414371 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b96c744c-10e9-4c0c-9ece-069c0c90e67a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.414474 4840 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.414591 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72l8s\" (UniqueName: \"kubernetes.io/projected/b96c744c-10e9-4c0c-9ece-069c0c90e67a-kube-api-access-72l8s\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.414684 4840 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b96c744c-10e9-4c0c-9ece-069c0c90e67a-logs\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.414744 4840 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b96c744c-10e9-4c0c-9ece-069c0c90e67a-public-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.414836 4840 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b96c744c-10e9-4c0c-9ece-069c0c90e67a-httpd-run\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.437495 4840 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.516860 4840 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.764113 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.826146 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c499ea9a-7047-493b-a529-1b5115119b25","Type":"ContainerDied","Data":"fab1facb1e72c8f3a2d0b0dd31e3b630edb5c37b802f0c642c83a48db6158c1e"} Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.826213 4840 scope.go:117] "RemoveContainer" containerID="b203574d1cba5c9043bdf32ea9de8523a03e3e11b59553bd30a8431865d24f1d" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.826176 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.829429 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b96c744c-10e9-4c0c-9ece-069c0c90e67a","Type":"ContainerDied","Data":"3418dbe2ddffdf60c07068d9f73484617a103469d7d8c7af4840e35d4d00c760"} Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.829497 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.915695 4840 scope.go:117] "RemoveContainer" containerID="03648621acfcfc360502f4b3af076777d7d5f8a4f30a001515922855a9ab1fe9" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.923701 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c499ea9a-7047-493b-a529-1b5115119b25-combined-ca-bundle\") pod \"c499ea9a-7047-493b-a529-1b5115119b25\" (UID: \"c499ea9a-7047-493b-a529-1b5115119b25\") " Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.923795 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c499ea9a-7047-493b-a529-1b5115119b25-httpd-run\") pod \"c499ea9a-7047-493b-a529-1b5115119b25\" (UID: \"c499ea9a-7047-493b-a529-1b5115119b25\") " Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.923864 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c499ea9a-7047-493b-a529-1b5115119b25-logs\") pod \"c499ea9a-7047-493b-a529-1b5115119b25\" (UID: \"c499ea9a-7047-493b-a529-1b5115119b25\") " Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.923886 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c499ea9a-7047-493b-a529-1b5115119b25-internal-tls-certs\") pod \"c499ea9a-7047-493b-a529-1b5115119b25\" (UID: \"c499ea9a-7047-493b-a529-1b5115119b25\") " Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.923972 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c499ea9a-7047-493b-a529-1b5115119b25-scripts\") pod \"c499ea9a-7047-493b-a529-1b5115119b25\" (UID: \"c499ea9a-7047-493b-a529-1b5115119b25\") " Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.923996 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hs4t6\" (UniqueName: \"kubernetes.io/projected/c499ea9a-7047-493b-a529-1b5115119b25-kube-api-access-hs4t6\") pod \"c499ea9a-7047-493b-a529-1b5115119b25\" (UID: \"c499ea9a-7047-493b-a529-1b5115119b25\") " Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.924027 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"c499ea9a-7047-493b-a529-1b5115119b25\" (UID: \"c499ea9a-7047-493b-a529-1b5115119b25\") " Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.924081 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c499ea9a-7047-493b-a529-1b5115119b25-config-data\") pod \"c499ea9a-7047-493b-a529-1b5115119b25\" (UID: \"c499ea9a-7047-493b-a529-1b5115119b25\") " Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.928354 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c499ea9a-7047-493b-a529-1b5115119b25-logs" (OuterVolumeSpecName: "logs") pod "c499ea9a-7047-493b-a529-1b5115119b25" (UID: "c499ea9a-7047-493b-a529-1b5115119b25"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.928457 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.929959 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "c499ea9a-7047-493b-a529-1b5115119b25" (UID: "c499ea9a-7047-493b-a529-1b5115119b25"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.930386 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c499ea9a-7047-493b-a529-1b5115119b25-scripts" (OuterVolumeSpecName: "scripts") pod "c499ea9a-7047-493b-a529-1b5115119b25" (UID: "c499ea9a-7047-493b-a529-1b5115119b25"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.933814 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c499ea9a-7047-493b-a529-1b5115119b25-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "c499ea9a-7047-493b-a529-1b5115119b25" (UID: "c499ea9a-7047-493b-a529-1b5115119b25"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.937933 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c499ea9a-7047-493b-a529-1b5115119b25-kube-api-access-hs4t6" (OuterVolumeSpecName: "kube-api-access-hs4t6") pod "c499ea9a-7047-493b-a529-1b5115119b25" (UID: "c499ea9a-7047-493b-a529-1b5115119b25"). InnerVolumeSpecName "kube-api-access-hs4t6". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.941482 4840 scope.go:117] "RemoveContainer" containerID="3c95288a14cc91765a03616025aa0e85a31db07cc1cc694550d048ce584c730b" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.941803 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.970047 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 14:14:06 crc kubenswrapper[4840]: E0930 14:14:06.970501 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b96c744c-10e9-4c0c-9ece-069c0c90e67a" containerName="glance-log" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.970514 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="b96c744c-10e9-4c0c-9ece-069c0c90e67a" containerName="glance-log" Sep 30 14:14:06 crc kubenswrapper[4840]: E0930 14:14:06.970579 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b96c744c-10e9-4c0c-9ece-069c0c90e67a" containerName="glance-httpd" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.970589 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="b96c744c-10e9-4c0c-9ece-069c0c90e67a" containerName="glance-httpd" Sep 30 14:14:06 crc kubenswrapper[4840]: E0930 14:14:06.970602 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12f02cd8-eb57-4557-b08a-1fc38f1f49fd" containerName="mariadb-database-create" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.970611 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="12f02cd8-eb57-4557-b08a-1fc38f1f49fd" containerName="mariadb-database-create" Sep 30 14:14:06 crc kubenswrapper[4840]: E0930 14:14:06.970624 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c499ea9a-7047-493b-a529-1b5115119b25" containerName="glance-log" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.970630 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="c499ea9a-7047-493b-a529-1b5115119b25" containerName="glance-log" Sep 30 14:14:06 crc kubenswrapper[4840]: E0930 14:14:06.970641 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="252dac07-3c44-4264-a488-cca59b98e516" containerName="mariadb-database-create" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.970646 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="252dac07-3c44-4264-a488-cca59b98e516" containerName="mariadb-database-create" Sep 30 14:14:06 crc kubenswrapper[4840]: E0930 14:14:06.970658 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c499ea9a-7047-493b-a529-1b5115119b25" containerName="glance-httpd" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.970663 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="c499ea9a-7047-493b-a529-1b5115119b25" containerName="glance-httpd" Sep 30 14:14:06 crc kubenswrapper[4840]: E0930 14:14:06.970676 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e00317b4-982d-45c8-b5d6-6b69e4d1d0a7" containerName="mariadb-database-create" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.970682 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="e00317b4-982d-45c8-b5d6-6b69e4d1d0a7" containerName="mariadb-database-create" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.970852 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="b96c744c-10e9-4c0c-9ece-069c0c90e67a" containerName="glance-log" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.970865 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="b96c744c-10e9-4c0c-9ece-069c0c90e67a" containerName="glance-httpd" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.970887 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="e00317b4-982d-45c8-b5d6-6b69e4d1d0a7" containerName="mariadb-database-create" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.970900 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="252dac07-3c44-4264-a488-cca59b98e516" containerName="mariadb-database-create" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.970913 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="c499ea9a-7047-493b-a529-1b5115119b25" containerName="glance-httpd" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.970921 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="c499ea9a-7047-493b-a529-1b5115119b25" containerName="glance-log" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.970932 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="12f02cd8-eb57-4557-b08a-1fc38f1f49fd" containerName="mariadb-database-create" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.972002 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.974242 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.974513 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.975079 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c499ea9a-7047-493b-a529-1b5115119b25-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c499ea9a-7047-493b-a529-1b5115119b25" (UID: "c499ea9a-7047-493b-a529-1b5115119b25"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.986200 4840 scope.go:117] "RemoveContainer" containerID="f19c53a63f714f8e0a84ed1833fdf7a816a89d7939ff5aad1ade067169e78a35" Sep 30 14:14:06 crc kubenswrapper[4840]: I0930 14:14:06.994145 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.000178 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c499ea9a-7047-493b-a529-1b5115119b25-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "c499ea9a-7047-493b-a529-1b5115119b25" (UID: "c499ea9a-7047-493b-a529-1b5115119b25"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.008653 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c499ea9a-7047-493b-a529-1b5115119b25-config-data" (OuterVolumeSpecName: "config-data") pod "c499ea9a-7047-493b-a529-1b5115119b25" (UID: "c499ea9a-7047-493b-a529-1b5115119b25"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.026161 4840 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c499ea9a-7047-493b-a529-1b5115119b25-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.026198 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hs4t6\" (UniqueName: \"kubernetes.io/projected/c499ea9a-7047-493b-a529-1b5115119b25-kube-api-access-hs4t6\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.026260 4840 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.026275 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c499ea9a-7047-493b-a529-1b5115119b25-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.026285 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c499ea9a-7047-493b-a529-1b5115119b25-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.026293 4840 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c499ea9a-7047-493b-a529-1b5115119b25-httpd-run\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.026302 4840 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c499ea9a-7047-493b-a529-1b5115119b25-logs\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.026309 4840 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c499ea9a-7047-493b-a529-1b5115119b25-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.045479 4840 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.127932 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1e73274-948e-4813-946c-0d98ce327e91-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b1e73274-948e-4813-946c-0d98ce327e91\") " pod="openstack/glance-default-external-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.128022 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1e73274-948e-4813-946c-0d98ce327e91-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"b1e73274-948e-4813-946c-0d98ce327e91\") " pod="openstack/glance-default-external-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.128050 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1e73274-948e-4813-946c-0d98ce327e91-config-data\") pod \"glance-default-external-api-0\" (UID: \"b1e73274-948e-4813-946c-0d98ce327e91\") " pod="openstack/glance-default-external-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.128090 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1e73274-948e-4813-946c-0d98ce327e91-scripts\") pod \"glance-default-external-api-0\" (UID: \"b1e73274-948e-4813-946c-0d98ce327e91\") " pod="openstack/glance-default-external-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.128205 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgnxw\" (UniqueName: \"kubernetes.io/projected/b1e73274-948e-4813-946c-0d98ce327e91-kube-api-access-sgnxw\") pod \"glance-default-external-api-0\" (UID: \"b1e73274-948e-4813-946c-0d98ce327e91\") " pod="openstack/glance-default-external-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.128266 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"b1e73274-948e-4813-946c-0d98ce327e91\") " pod="openstack/glance-default-external-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.128359 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1e73274-948e-4813-946c-0d98ce327e91-logs\") pod \"glance-default-external-api-0\" (UID: \"b1e73274-948e-4813-946c-0d98ce327e91\") " pod="openstack/glance-default-external-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.128423 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b1e73274-948e-4813-946c-0d98ce327e91-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b1e73274-948e-4813-946c-0d98ce327e91\") " pod="openstack/glance-default-external-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.128583 4840 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.157866 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.164839 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.181777 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.184725 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.187364 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.192881 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.215153 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.230122 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgnxw\" (UniqueName: \"kubernetes.io/projected/b1e73274-948e-4813-946c-0d98ce327e91-kube-api-access-sgnxw\") pod \"glance-default-external-api-0\" (UID: \"b1e73274-948e-4813-946c-0d98ce327e91\") " pod="openstack/glance-default-external-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.230442 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"c58eeb6f-560e-4abf-b604-65a3d2819ac5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.230559 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"b1e73274-948e-4813-946c-0d98ce327e91\") " pod="openstack/glance-default-external-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.230702 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c58eeb6f-560e-4abf-b604-65a3d2819ac5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c58eeb6f-560e-4abf-b604-65a3d2819ac5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.230818 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1e73274-948e-4813-946c-0d98ce327e91-logs\") pod \"glance-default-external-api-0\" (UID: \"b1e73274-948e-4813-946c-0d98ce327e91\") " pod="openstack/glance-default-external-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.230943 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b1e73274-948e-4813-946c-0d98ce327e91-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b1e73274-948e-4813-946c-0d98ce327e91\") " pod="openstack/glance-default-external-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.231050 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c58eeb6f-560e-4abf-b604-65a3d2819ac5-logs\") pod \"glance-default-internal-api-0\" (UID: \"c58eeb6f-560e-4abf-b604-65a3d2819ac5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.231147 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c58eeb6f-560e-4abf-b604-65a3d2819ac5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c58eeb6f-560e-4abf-b604-65a3d2819ac5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.231262 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c58eeb6f-560e-4abf-b604-65a3d2819ac5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c58eeb6f-560e-4abf-b604-65a3d2819ac5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.230728 4840 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"b1e73274-948e-4813-946c-0d98ce327e91\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.231460 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1e73274-948e-4813-946c-0d98ce327e91-logs\") pod \"glance-default-external-api-0\" (UID: \"b1e73274-948e-4813-946c-0d98ce327e91\") " pod="openstack/glance-default-external-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.231484 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b1e73274-948e-4813-946c-0d98ce327e91-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b1e73274-948e-4813-946c-0d98ce327e91\") " pod="openstack/glance-default-external-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.231665 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drgzh\" (UniqueName: \"kubernetes.io/projected/c58eeb6f-560e-4abf-b604-65a3d2819ac5-kube-api-access-drgzh\") pod \"glance-default-internal-api-0\" (UID: \"c58eeb6f-560e-4abf-b604-65a3d2819ac5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.231826 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1e73274-948e-4813-946c-0d98ce327e91-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b1e73274-948e-4813-946c-0d98ce327e91\") " pod="openstack/glance-default-external-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.231950 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1e73274-948e-4813-946c-0d98ce327e91-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"b1e73274-948e-4813-946c-0d98ce327e91\") " pod="openstack/glance-default-external-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.231978 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1e73274-948e-4813-946c-0d98ce327e91-config-data\") pod \"glance-default-external-api-0\" (UID: \"b1e73274-948e-4813-946c-0d98ce327e91\") " pod="openstack/glance-default-external-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.232027 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c58eeb6f-560e-4abf-b604-65a3d2819ac5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c58eeb6f-560e-4abf-b604-65a3d2819ac5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.232079 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1e73274-948e-4813-946c-0d98ce327e91-scripts\") pod \"glance-default-external-api-0\" (UID: \"b1e73274-948e-4813-946c-0d98ce327e91\") " pod="openstack/glance-default-external-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.232103 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58eeb6f-560e-4abf-b604-65a3d2819ac5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c58eeb6f-560e-4abf-b604-65a3d2819ac5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.236275 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1e73274-948e-4813-946c-0d98ce327e91-scripts\") pod \"glance-default-external-api-0\" (UID: \"b1e73274-948e-4813-946c-0d98ce327e91\") " pod="openstack/glance-default-external-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.236903 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1e73274-948e-4813-946c-0d98ce327e91-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b1e73274-948e-4813-946c-0d98ce327e91\") " pod="openstack/glance-default-external-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.237713 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1e73274-948e-4813-946c-0d98ce327e91-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"b1e73274-948e-4813-946c-0d98ce327e91\") " pod="openstack/glance-default-external-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.238483 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1e73274-948e-4813-946c-0d98ce327e91-config-data\") pod \"glance-default-external-api-0\" (UID: \"b1e73274-948e-4813-946c-0d98ce327e91\") " pod="openstack/glance-default-external-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.247434 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgnxw\" (UniqueName: \"kubernetes.io/projected/b1e73274-948e-4813-946c-0d98ce327e91-kube-api-access-sgnxw\") pod \"glance-default-external-api-0\" (UID: \"b1e73274-948e-4813-946c-0d98ce327e91\") " pod="openstack/glance-default-external-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.258721 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"b1e73274-948e-4813-946c-0d98ce327e91\") " pod="openstack/glance-default-external-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.299862 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.333488 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c58eeb6f-560e-4abf-b604-65a3d2819ac5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c58eeb6f-560e-4abf-b604-65a3d2819ac5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.333530 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58eeb6f-560e-4abf-b604-65a3d2819ac5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c58eeb6f-560e-4abf-b604-65a3d2819ac5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.333573 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"c58eeb6f-560e-4abf-b604-65a3d2819ac5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.333618 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c58eeb6f-560e-4abf-b604-65a3d2819ac5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c58eeb6f-560e-4abf-b604-65a3d2819ac5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.333667 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c58eeb6f-560e-4abf-b604-65a3d2819ac5-logs\") pod \"glance-default-internal-api-0\" (UID: \"c58eeb6f-560e-4abf-b604-65a3d2819ac5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.333692 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c58eeb6f-560e-4abf-b604-65a3d2819ac5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c58eeb6f-560e-4abf-b604-65a3d2819ac5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.333715 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c58eeb6f-560e-4abf-b604-65a3d2819ac5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c58eeb6f-560e-4abf-b604-65a3d2819ac5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.333765 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drgzh\" (UniqueName: \"kubernetes.io/projected/c58eeb6f-560e-4abf-b604-65a3d2819ac5-kube-api-access-drgzh\") pod \"glance-default-internal-api-0\" (UID: \"c58eeb6f-560e-4abf-b604-65a3d2819ac5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.333808 4840 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"c58eeb6f-560e-4abf-b604-65a3d2819ac5\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-internal-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.334836 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c58eeb6f-560e-4abf-b604-65a3d2819ac5-logs\") pod \"glance-default-internal-api-0\" (UID: \"c58eeb6f-560e-4abf-b604-65a3d2819ac5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.335052 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c58eeb6f-560e-4abf-b604-65a3d2819ac5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c58eeb6f-560e-4abf-b604-65a3d2819ac5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.337048 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c58eeb6f-560e-4abf-b604-65a3d2819ac5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c58eeb6f-560e-4abf-b604-65a3d2819ac5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.338976 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58eeb6f-560e-4abf-b604-65a3d2819ac5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c58eeb6f-560e-4abf-b604-65a3d2819ac5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.340988 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c58eeb6f-560e-4abf-b604-65a3d2819ac5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c58eeb6f-560e-4abf-b604-65a3d2819ac5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.341364 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c58eeb6f-560e-4abf-b604-65a3d2819ac5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c58eeb6f-560e-4abf-b604-65a3d2819ac5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.352624 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drgzh\" (UniqueName: \"kubernetes.io/projected/c58eeb6f-560e-4abf-b604-65a3d2819ac5-kube-api-access-drgzh\") pod \"glance-default-internal-api-0\" (UID: \"c58eeb6f-560e-4abf-b604-65a3d2819ac5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.367460 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"c58eeb6f-560e-4abf-b604-65a3d2819ac5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.511950 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.634658 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-60a3-account-create-qjn85"] Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.636096 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-60a3-account-create-qjn85" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.639583 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.643793 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-60a3-account-create-qjn85"] Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.740342 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlp28\" (UniqueName: \"kubernetes.io/projected/3fe3cf3b-8d33-4689-86c9-3331fdcc6fac-kube-api-access-nlp28\") pod \"nova-api-60a3-account-create-qjn85\" (UID: \"3fe3cf3b-8d33-4689-86c9-3331fdcc6fac\") " pod="openstack/nova-api-60a3-account-create-qjn85" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.813657 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-6731-account-create-v45sj"] Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.814792 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-6731-account-create-v45sj" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.816532 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.830944 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-6731-account-create-v45sj"] Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.841787 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6fjv\" (UniqueName: \"kubernetes.io/projected/4c2fd404-193d-4a27-92c9-395b2c0fcc02-kube-api-access-m6fjv\") pod \"nova-cell1-6731-account-create-v45sj\" (UID: \"4c2fd404-193d-4a27-92c9-395b2c0fcc02\") " pod="openstack/nova-cell1-6731-account-create-v45sj" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.841864 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlp28\" (UniqueName: \"kubernetes.io/projected/3fe3cf3b-8d33-4689-86c9-3331fdcc6fac-kube-api-access-nlp28\") pod \"nova-api-60a3-account-create-qjn85\" (UID: \"3fe3cf3b-8d33-4689-86c9-3331fdcc6fac\") " pod="openstack/nova-api-60a3-account-create-qjn85" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.861404 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlp28\" (UniqueName: \"kubernetes.io/projected/3fe3cf3b-8d33-4689-86c9-3331fdcc6fac-kube-api-access-nlp28\") pod \"nova-api-60a3-account-create-qjn85\" (UID: \"3fe3cf3b-8d33-4689-86c9-3331fdcc6fac\") " pod="openstack/nova-api-60a3-account-create-qjn85" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.943899 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6fjv\" (UniqueName: \"kubernetes.io/projected/4c2fd404-193d-4a27-92c9-395b2c0fcc02-kube-api-access-m6fjv\") pod \"nova-cell1-6731-account-create-v45sj\" (UID: \"4c2fd404-193d-4a27-92c9-395b2c0fcc02\") " pod="openstack/nova-cell1-6731-account-create-v45sj" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.958139 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-60a3-account-create-qjn85" Sep 30 14:14:07 crc kubenswrapper[4840]: I0930 14:14:07.963339 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6fjv\" (UniqueName: \"kubernetes.io/projected/4c2fd404-193d-4a27-92c9-395b2c0fcc02-kube-api-access-m6fjv\") pod \"nova-cell1-6731-account-create-v45sj\" (UID: \"4c2fd404-193d-4a27-92c9-395b2c0fcc02\") " pod="openstack/nova-cell1-6731-account-create-v45sj" Sep 30 14:14:08 crc kubenswrapper[4840]: I0930 14:14:08.133514 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b96c744c-10e9-4c0c-9ece-069c0c90e67a" path="/var/lib/kubelet/pods/b96c744c-10e9-4c0c-9ece-069c0c90e67a/volumes" Sep 30 14:14:08 crc kubenswrapper[4840]: I0930 14:14:08.135514 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c499ea9a-7047-493b-a529-1b5115119b25" path="/var/lib/kubelet/pods/c499ea9a-7047-493b-a529-1b5115119b25/volumes" Sep 30 14:14:08 crc kubenswrapper[4840]: I0930 14:14:08.138012 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-6731-account-create-v45sj" Sep 30 14:14:08 crc kubenswrapper[4840]: I0930 14:14:08.564396 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 14:14:08 crc kubenswrapper[4840]: I0930 14:14:08.594738 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5595d79c68-dgsth" podUID="5e9b87b2-4836-4857-8514-b2dcfe4de1f5" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Sep 30 14:14:08 crc kubenswrapper[4840]: I0930 14:14:08.594853 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5595d79c68-dgsth" Sep 30 14:14:08 crc kubenswrapper[4840]: W0930 14:14:08.908374 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1e73274_948e_4813_946c_0d98ce327e91.slice/crio-d137552b4f41ee25247420d306a03047c36399f0ba35db66a8fcfba23760549a WatchSource:0}: Error finding container d137552b4f41ee25247420d306a03047c36399f0ba35db66a8fcfba23760549a: Status 404 returned error can't find the container with id d137552b4f41ee25247420d306a03047c36399f0ba35db66a8fcfba23760549a Sep 30 14:14:09 crc kubenswrapper[4840]: I0930 14:14:09.436676 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-60a3-account-create-qjn85"] Sep 30 14:14:09 crc kubenswrapper[4840]: I0930 14:14:09.586143 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-6731-account-create-v45sj"] Sep 30 14:14:09 crc kubenswrapper[4840]: I0930 14:14:09.598202 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:14:09 crc kubenswrapper[4840]: W0930 14:14:09.616445 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc58eeb6f_560e_4abf_b604_65a3d2819ac5.slice/crio-46ad69b3c9eb00c1f28a99228bb3b10c7e0ee089a821385e74c9570e83f72293 WatchSource:0}: Error finding container 46ad69b3c9eb00c1f28a99228bb3b10c7e0ee089a821385e74c9570e83f72293: Status 404 returned error can't find the container with id 46ad69b3c9eb00c1f28a99228bb3b10c7e0ee089a821385e74c9570e83f72293 Sep 30 14:14:09 crc kubenswrapper[4840]: I0930 14:14:09.872287 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c58eeb6f-560e-4abf-b604-65a3d2819ac5","Type":"ContainerStarted","Data":"46ad69b3c9eb00c1f28a99228bb3b10c7e0ee089a821385e74c9570e83f72293"} Sep 30 14:14:09 crc kubenswrapper[4840]: I0930 14:14:09.875097 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b1e73274-948e-4813-946c-0d98ce327e91","Type":"ContainerStarted","Data":"19f63dba1372b5e8583371523abdbd6878faa503463bdde7a8631b64b93cd30a"} Sep 30 14:14:09 crc kubenswrapper[4840]: I0930 14:14:09.875126 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b1e73274-948e-4813-946c-0d98ce327e91","Type":"ContainerStarted","Data":"d137552b4f41ee25247420d306a03047c36399f0ba35db66a8fcfba23760549a"} Sep 30 14:14:09 crc kubenswrapper[4840]: I0930 14:14:09.876813 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-6731-account-create-v45sj" event={"ID":"4c2fd404-193d-4a27-92c9-395b2c0fcc02","Type":"ContainerStarted","Data":"5a980ce6f702025cb46efebfcd6baf9b79b79afd7f53a1b121cd10497d1b3ad1"} Sep 30 14:14:09 crc kubenswrapper[4840]: I0930 14:14:09.876836 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-6731-account-create-v45sj" event={"ID":"4c2fd404-193d-4a27-92c9-395b2c0fcc02","Type":"ContainerStarted","Data":"56f01fb854b42d48669a98c6fd5b387097cd8ca67102329c1ea9cbd9c4ea1677"} Sep 30 14:14:09 crc kubenswrapper[4840]: I0930 14:14:09.879116 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-60a3-account-create-qjn85" event={"ID":"3fe3cf3b-8d33-4689-86c9-3331fdcc6fac","Type":"ContainerStarted","Data":"f513d577284ff2ffc61e31d9e5458e15e06f530cc0498bd6d4e92a4378473124"} Sep 30 14:14:09 crc kubenswrapper[4840]: I0930 14:14:09.879140 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-60a3-account-create-qjn85" event={"ID":"3fe3cf3b-8d33-4689-86c9-3331fdcc6fac","Type":"ContainerStarted","Data":"0206ad3fd8b782cf94a0a075bfde58c5678c1cc6c48a6463195c8a345c0e9af8"} Sep 30 14:14:09 crc kubenswrapper[4840]: I0930 14:14:09.882373 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf","Type":"ContainerStarted","Data":"e289d0fde27d2ec83cbdf6dd13368f1d42bf91a3992499cf862f7683b6687fc3"} Sep 30 14:14:10 crc kubenswrapper[4840]: I0930 14:14:10.897119 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b1e73274-948e-4813-946c-0d98ce327e91","Type":"ContainerStarted","Data":"e7a4980b3ddd001e1724e20cbcbf14628af46debd4cc1aef856042fbbabed8f1"} Sep 30 14:14:10 crc kubenswrapper[4840]: I0930 14:14:10.905589 4840 generic.go:334] "Generic (PLEG): container finished" podID="4c2fd404-193d-4a27-92c9-395b2c0fcc02" containerID="5a980ce6f702025cb46efebfcd6baf9b79b79afd7f53a1b121cd10497d1b3ad1" exitCode=0 Sep 30 14:14:10 crc kubenswrapper[4840]: I0930 14:14:10.905718 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-6731-account-create-v45sj" event={"ID":"4c2fd404-193d-4a27-92c9-395b2c0fcc02","Type":"ContainerDied","Data":"5a980ce6f702025cb46efebfcd6baf9b79b79afd7f53a1b121cd10497d1b3ad1"} Sep 30 14:14:10 crc kubenswrapper[4840]: I0930 14:14:10.907050 4840 generic.go:334] "Generic (PLEG): container finished" podID="3fe3cf3b-8d33-4689-86c9-3331fdcc6fac" containerID="f513d577284ff2ffc61e31d9e5458e15e06f530cc0498bd6d4e92a4378473124" exitCode=0 Sep 30 14:14:10 crc kubenswrapper[4840]: I0930 14:14:10.907123 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-60a3-account-create-qjn85" event={"ID":"3fe3cf3b-8d33-4689-86c9-3331fdcc6fac","Type":"ContainerDied","Data":"f513d577284ff2ffc61e31d9e5458e15e06f530cc0498bd6d4e92a4378473124"} Sep 30 14:14:10 crc kubenswrapper[4840]: I0930 14:14:10.908136 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c58eeb6f-560e-4abf-b604-65a3d2819ac5","Type":"ContainerStarted","Data":"a2de00e2749a1c1027a6065c090364579bfedf6b31638bcb30a89e48f02538c7"} Sep 30 14:14:10 crc kubenswrapper[4840]: I0930 14:14:10.939837 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.939818089 podStartE2EDuration="4.939818089s" podCreationTimestamp="2025-09-30 14:14:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:14:10.936935494 +0000 UTC m=+1079.566021917" watchObservedRunningTime="2025-09-30 14:14:10.939818089 +0000 UTC m=+1079.568904512" Sep 30 14:14:11 crc kubenswrapper[4840]: I0930 14:14:11.776973 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5d586fc9-9z92d" Sep 30 14:14:11 crc kubenswrapper[4840]: I0930 14:14:11.779845 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5d586fc9-9z92d" Sep 30 14:14:11 crc kubenswrapper[4840]: I0930 14:14:11.945725 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c58eeb6f-560e-4abf-b604-65a3d2819ac5","Type":"ContainerStarted","Data":"72b2c5b1aa7ad06523c923b8274f68da44ec66ebc64b0d442032e5d0c74490e1"} Sep 30 14:14:11 crc kubenswrapper[4840]: I0930 14:14:11.953148 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf","Type":"ContainerStarted","Data":"2108585a2ecf8b5b32dac6ce916c88a80d75e7632fc6058de645c91402dcc389"} Sep 30 14:14:11 crc kubenswrapper[4840]: I0930 14:14:11.987006 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.986981298 podStartE2EDuration="4.986981298s" podCreationTimestamp="2025-09-30 14:14:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:14:11.966058062 +0000 UTC m=+1080.595144485" watchObservedRunningTime="2025-09-30 14:14:11.986981298 +0000 UTC m=+1080.616067721" Sep 30 14:14:12 crc kubenswrapper[4840]: I0930 14:14:12.389303 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-60a3-account-create-qjn85" Sep 30 14:14:12 crc kubenswrapper[4840]: I0930 14:14:12.396831 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-6731-account-create-v45sj" Sep 30 14:14:12 crc kubenswrapper[4840]: I0930 14:14:12.531127 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nlp28\" (UniqueName: \"kubernetes.io/projected/3fe3cf3b-8d33-4689-86c9-3331fdcc6fac-kube-api-access-nlp28\") pod \"3fe3cf3b-8d33-4689-86c9-3331fdcc6fac\" (UID: \"3fe3cf3b-8d33-4689-86c9-3331fdcc6fac\") " Sep 30 14:14:12 crc kubenswrapper[4840]: I0930 14:14:12.531206 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6fjv\" (UniqueName: \"kubernetes.io/projected/4c2fd404-193d-4a27-92c9-395b2c0fcc02-kube-api-access-m6fjv\") pod \"4c2fd404-193d-4a27-92c9-395b2c0fcc02\" (UID: \"4c2fd404-193d-4a27-92c9-395b2c0fcc02\") " Sep 30 14:14:12 crc kubenswrapper[4840]: I0930 14:14:12.537910 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c2fd404-193d-4a27-92c9-395b2c0fcc02-kube-api-access-m6fjv" (OuterVolumeSpecName: "kube-api-access-m6fjv") pod "4c2fd404-193d-4a27-92c9-395b2c0fcc02" (UID: "4c2fd404-193d-4a27-92c9-395b2c0fcc02"). InnerVolumeSpecName "kube-api-access-m6fjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:14:12 crc kubenswrapper[4840]: I0930 14:14:12.544226 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fe3cf3b-8d33-4689-86c9-3331fdcc6fac-kube-api-access-nlp28" (OuterVolumeSpecName: "kube-api-access-nlp28") pod "3fe3cf3b-8d33-4689-86c9-3331fdcc6fac" (UID: "3fe3cf3b-8d33-4689-86c9-3331fdcc6fac"). InnerVolumeSpecName "kube-api-access-nlp28". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:14:12 crc kubenswrapper[4840]: I0930 14:14:12.632800 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nlp28\" (UniqueName: \"kubernetes.io/projected/3fe3cf3b-8d33-4689-86c9-3331fdcc6fac-kube-api-access-nlp28\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:12 crc kubenswrapper[4840]: I0930 14:14:12.632827 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6fjv\" (UniqueName: \"kubernetes.io/projected/4c2fd404-193d-4a27-92c9-395b2c0fcc02-kube-api-access-m6fjv\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:12 crc kubenswrapper[4840]: I0930 14:14:12.962790 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-6731-account-create-v45sj" event={"ID":"4c2fd404-193d-4a27-92c9-395b2c0fcc02","Type":"ContainerDied","Data":"56f01fb854b42d48669a98c6fd5b387097cd8ca67102329c1ea9cbd9c4ea1677"} Sep 30 14:14:12 crc kubenswrapper[4840]: I0930 14:14:12.963508 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="56f01fb854b42d48669a98c6fd5b387097cd8ca67102329c1ea9cbd9c4ea1677" Sep 30 14:14:12 crc kubenswrapper[4840]: I0930 14:14:12.962819 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-6731-account-create-v45sj" Sep 30 14:14:12 crc kubenswrapper[4840]: I0930 14:14:12.964185 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-60a3-account-create-qjn85" Sep 30 14:14:12 crc kubenswrapper[4840]: I0930 14:14:12.964182 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-60a3-account-create-qjn85" event={"ID":"3fe3cf3b-8d33-4689-86c9-3331fdcc6fac","Type":"ContainerDied","Data":"0206ad3fd8b782cf94a0a075bfde58c5678c1cc6c48a6463195c8a345c0e9af8"} Sep 30 14:14:12 crc kubenswrapper[4840]: I0930 14:14:12.964345 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0206ad3fd8b782cf94a0a075bfde58c5678c1cc6c48a6463195c8a345c0e9af8" Sep 30 14:14:13 crc kubenswrapper[4840]: I0930 14:14:13.758217 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5595d79c68-dgsth" Sep 30 14:14:13 crc kubenswrapper[4840]: I0930 14:14:13.855940 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-scripts\") pod \"5e9b87b2-4836-4857-8514-b2dcfe4de1f5\" (UID: \"5e9b87b2-4836-4857-8514-b2dcfe4de1f5\") " Sep 30 14:14:13 crc kubenswrapper[4840]: I0930 14:14:13.856042 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-logs\") pod \"5e9b87b2-4836-4857-8514-b2dcfe4de1f5\" (UID: \"5e9b87b2-4836-4857-8514-b2dcfe4de1f5\") " Sep 30 14:14:13 crc kubenswrapper[4840]: I0930 14:14:13.856102 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-config-data\") pod \"5e9b87b2-4836-4857-8514-b2dcfe4de1f5\" (UID: \"5e9b87b2-4836-4857-8514-b2dcfe4de1f5\") " Sep 30 14:14:13 crc kubenswrapper[4840]: I0930 14:14:13.856230 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-combined-ca-bundle\") pod \"5e9b87b2-4836-4857-8514-b2dcfe4de1f5\" (UID: \"5e9b87b2-4836-4857-8514-b2dcfe4de1f5\") " Sep 30 14:14:13 crc kubenswrapper[4840]: I0930 14:14:13.856290 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmn5v\" (UniqueName: \"kubernetes.io/projected/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-kube-api-access-xmn5v\") pod \"5e9b87b2-4836-4857-8514-b2dcfe4de1f5\" (UID: \"5e9b87b2-4836-4857-8514-b2dcfe4de1f5\") " Sep 30 14:14:13 crc kubenswrapper[4840]: I0930 14:14:13.856356 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-horizon-tls-certs\") pod \"5e9b87b2-4836-4857-8514-b2dcfe4de1f5\" (UID: \"5e9b87b2-4836-4857-8514-b2dcfe4de1f5\") " Sep 30 14:14:13 crc kubenswrapper[4840]: I0930 14:14:13.856390 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-horizon-secret-key\") pod \"5e9b87b2-4836-4857-8514-b2dcfe4de1f5\" (UID: \"5e9b87b2-4836-4857-8514-b2dcfe4de1f5\") " Sep 30 14:14:13 crc kubenswrapper[4840]: I0930 14:14:13.857418 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-logs" (OuterVolumeSpecName: "logs") pod "5e9b87b2-4836-4857-8514-b2dcfe4de1f5" (UID: "5e9b87b2-4836-4857-8514-b2dcfe4de1f5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:14:13 crc kubenswrapper[4840]: I0930 14:14:13.864686 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "5e9b87b2-4836-4857-8514-b2dcfe4de1f5" (UID: "5e9b87b2-4836-4857-8514-b2dcfe4de1f5"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:14:13 crc kubenswrapper[4840]: I0930 14:14:13.864734 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-kube-api-access-xmn5v" (OuterVolumeSpecName: "kube-api-access-xmn5v") pod "5e9b87b2-4836-4857-8514-b2dcfe4de1f5" (UID: "5e9b87b2-4836-4857-8514-b2dcfe4de1f5"). InnerVolumeSpecName "kube-api-access-xmn5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:14:13 crc kubenswrapper[4840]: I0930 14:14:13.892778 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-config-data" (OuterVolumeSpecName: "config-data") pod "5e9b87b2-4836-4857-8514-b2dcfe4de1f5" (UID: "5e9b87b2-4836-4857-8514-b2dcfe4de1f5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:14:13 crc kubenswrapper[4840]: I0930 14:14:13.896684 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5e9b87b2-4836-4857-8514-b2dcfe4de1f5" (UID: "5e9b87b2-4836-4857-8514-b2dcfe4de1f5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:14:13 crc kubenswrapper[4840]: I0930 14:14:13.896866 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-scripts" (OuterVolumeSpecName: "scripts") pod "5e9b87b2-4836-4857-8514-b2dcfe4de1f5" (UID: "5e9b87b2-4836-4857-8514-b2dcfe4de1f5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:14:13 crc kubenswrapper[4840]: I0930 14:14:13.915610 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "5e9b87b2-4836-4857-8514-b2dcfe4de1f5" (UID: "5e9b87b2-4836-4857-8514-b2dcfe4de1f5"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:14:13 crc kubenswrapper[4840]: I0930 14:14:13.958814 4840 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:13 crc kubenswrapper[4840]: I0930 14:14:13.958843 4840 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:13 crc kubenswrapper[4840]: I0930 14:14:13.958857 4840 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:13 crc kubenswrapper[4840]: I0930 14:14:13.958865 4840 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-logs\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:13 crc kubenswrapper[4840]: I0930 14:14:13.958873 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:13 crc kubenswrapper[4840]: I0930 14:14:13.958880 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:13 crc kubenswrapper[4840]: I0930 14:14:13.958889 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmn5v\" (UniqueName: \"kubernetes.io/projected/5e9b87b2-4836-4857-8514-b2dcfe4de1f5-kube-api-access-xmn5v\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:13 crc kubenswrapper[4840]: I0930 14:14:13.978785 4840 generic.go:334] "Generic (PLEG): container finished" podID="5e9b87b2-4836-4857-8514-b2dcfe4de1f5" containerID="019b8bf8dc37878bad1354f6206205178cb6737805a6453683b754b0a592f5d1" exitCode=137 Sep 30 14:14:13 crc kubenswrapper[4840]: I0930 14:14:13.978827 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5595d79c68-dgsth" event={"ID":"5e9b87b2-4836-4857-8514-b2dcfe4de1f5","Type":"ContainerDied","Data":"019b8bf8dc37878bad1354f6206205178cb6737805a6453683b754b0a592f5d1"} Sep 30 14:14:13 crc kubenswrapper[4840]: I0930 14:14:13.978852 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5595d79c68-dgsth" event={"ID":"5e9b87b2-4836-4857-8514-b2dcfe4de1f5","Type":"ContainerDied","Data":"9b0c0ec11411cdf62ef15443f9485d30baa67c39b99c6937c92ff75d56b07dd0"} Sep 30 14:14:13 crc kubenswrapper[4840]: I0930 14:14:13.978868 4840 scope.go:117] "RemoveContainer" containerID="728e90849e942f659a1349a083627796a9122008fae88346ec1107d99eefbab1" Sep 30 14:14:13 crc kubenswrapper[4840]: I0930 14:14:13.979017 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5595d79c68-dgsth" Sep 30 14:14:14 crc kubenswrapper[4840]: I0930 14:14:14.015028 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5595d79c68-dgsth"] Sep 30 14:14:14 crc kubenswrapper[4840]: I0930 14:14:14.022384 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5595d79c68-dgsth"] Sep 30 14:14:14 crc kubenswrapper[4840]: I0930 14:14:14.129447 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e9b87b2-4836-4857-8514-b2dcfe4de1f5" path="/var/lib/kubelet/pods/5e9b87b2-4836-4857-8514-b2dcfe4de1f5/volumes" Sep 30 14:14:14 crc kubenswrapper[4840]: I0930 14:14:14.205411 4840 scope.go:117] "RemoveContainer" containerID="019b8bf8dc37878bad1354f6206205178cb6737805a6453683b754b0a592f5d1" Sep 30 14:14:14 crc kubenswrapper[4840]: I0930 14:14:14.404143 4840 scope.go:117] "RemoveContainer" containerID="728e90849e942f659a1349a083627796a9122008fae88346ec1107d99eefbab1" Sep 30 14:14:14 crc kubenswrapper[4840]: E0930 14:14:14.406616 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"728e90849e942f659a1349a083627796a9122008fae88346ec1107d99eefbab1\": container with ID starting with 728e90849e942f659a1349a083627796a9122008fae88346ec1107d99eefbab1 not found: ID does not exist" containerID="728e90849e942f659a1349a083627796a9122008fae88346ec1107d99eefbab1" Sep 30 14:14:14 crc kubenswrapper[4840]: I0930 14:14:14.406660 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"728e90849e942f659a1349a083627796a9122008fae88346ec1107d99eefbab1"} err="failed to get container status \"728e90849e942f659a1349a083627796a9122008fae88346ec1107d99eefbab1\": rpc error: code = NotFound desc = could not find container \"728e90849e942f659a1349a083627796a9122008fae88346ec1107d99eefbab1\": container with ID starting with 728e90849e942f659a1349a083627796a9122008fae88346ec1107d99eefbab1 not found: ID does not exist" Sep 30 14:14:14 crc kubenswrapper[4840]: I0930 14:14:14.406690 4840 scope.go:117] "RemoveContainer" containerID="019b8bf8dc37878bad1354f6206205178cb6737805a6453683b754b0a592f5d1" Sep 30 14:14:14 crc kubenswrapper[4840]: E0930 14:14:14.407291 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"019b8bf8dc37878bad1354f6206205178cb6737805a6453683b754b0a592f5d1\": container with ID starting with 019b8bf8dc37878bad1354f6206205178cb6737805a6453683b754b0a592f5d1 not found: ID does not exist" containerID="019b8bf8dc37878bad1354f6206205178cb6737805a6453683b754b0a592f5d1" Sep 30 14:14:14 crc kubenswrapper[4840]: I0930 14:14:14.407346 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"019b8bf8dc37878bad1354f6206205178cb6737805a6453683b754b0a592f5d1"} err="failed to get container status \"019b8bf8dc37878bad1354f6206205178cb6737805a6453683b754b0a592f5d1\": rpc error: code = NotFound desc = could not find container \"019b8bf8dc37878bad1354f6206205178cb6737805a6453683b754b0a592f5d1\": container with ID starting with 019b8bf8dc37878bad1354f6206205178cb6737805a6453683b754b0a592f5d1 not found: ID does not exist" Sep 30 14:14:14 crc kubenswrapper[4840]: I0930 14:14:14.989598 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf","Type":"ContainerStarted","Data":"63143e1a11e43c6259a3b43a048b0916e4859906600c2c4cc3f3b2db73fff7c4"} Sep 30 14:14:14 crc kubenswrapper[4840]: I0930 14:14:14.989670 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4aa3849f-5f45-4c05-ac5a-dc8472d5acaf" containerName="ceilometer-central-agent" containerID="cri-o://cfcaf10f26382eada414852d9cab5083c54f9e5f2466f123779ce16711278cf9" gracePeriod=30 Sep 30 14:14:14 crc kubenswrapper[4840]: I0930 14:14:14.989722 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 30 14:14:14 crc kubenswrapper[4840]: I0930 14:14:14.989774 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4aa3849f-5f45-4c05-ac5a-dc8472d5acaf" containerName="proxy-httpd" containerID="cri-o://63143e1a11e43c6259a3b43a048b0916e4859906600c2c4cc3f3b2db73fff7c4" gracePeriod=30 Sep 30 14:14:14 crc kubenswrapper[4840]: I0930 14:14:14.989819 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4aa3849f-5f45-4c05-ac5a-dc8472d5acaf" containerName="sg-core" containerID="cri-o://2108585a2ecf8b5b32dac6ce916c88a80d75e7632fc6058de645c91402dcc389" gracePeriod=30 Sep 30 14:14:14 crc kubenswrapper[4840]: I0930 14:14:14.989851 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4aa3849f-5f45-4c05-ac5a-dc8472d5acaf" containerName="ceilometer-notification-agent" containerID="cri-o://e289d0fde27d2ec83cbdf6dd13368f1d42bf91a3992499cf862f7683b6687fc3" gracePeriod=30 Sep 30 14:14:15 crc kubenswrapper[4840]: I0930 14:14:15.015541 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.252139421 podStartE2EDuration="14.015519996s" podCreationTimestamp="2025-09-30 14:14:01 +0000 UTC" firstStartedPulling="2025-09-30 14:14:02.641028724 +0000 UTC m=+1071.270115137" lastFinishedPulling="2025-09-30 14:14:14.404409289 +0000 UTC m=+1083.033495712" observedRunningTime="2025-09-30 14:14:15.010772702 +0000 UTC m=+1083.639859125" watchObservedRunningTime="2025-09-30 14:14:15.015519996 +0000 UTC m=+1083.644606419" Sep 30 14:14:16 crc kubenswrapper[4840]: I0930 14:14:16.005600 4840 generic.go:334] "Generic (PLEG): container finished" podID="4aa3849f-5f45-4c05-ac5a-dc8472d5acaf" containerID="63143e1a11e43c6259a3b43a048b0916e4859906600c2c4cc3f3b2db73fff7c4" exitCode=0 Sep 30 14:14:16 crc kubenswrapper[4840]: I0930 14:14:16.005889 4840 generic.go:334] "Generic (PLEG): container finished" podID="4aa3849f-5f45-4c05-ac5a-dc8472d5acaf" containerID="2108585a2ecf8b5b32dac6ce916c88a80d75e7632fc6058de645c91402dcc389" exitCode=2 Sep 30 14:14:16 crc kubenswrapper[4840]: I0930 14:14:16.005896 4840 generic.go:334] "Generic (PLEG): container finished" podID="4aa3849f-5f45-4c05-ac5a-dc8472d5acaf" containerID="e289d0fde27d2ec83cbdf6dd13368f1d42bf91a3992499cf862f7683b6687fc3" exitCode=0 Sep 30 14:14:16 crc kubenswrapper[4840]: I0930 14:14:16.005655 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf","Type":"ContainerDied","Data":"63143e1a11e43c6259a3b43a048b0916e4859906600c2c4cc3f3b2db73fff7c4"} Sep 30 14:14:16 crc kubenswrapper[4840]: I0930 14:14:16.005924 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf","Type":"ContainerDied","Data":"2108585a2ecf8b5b32dac6ce916c88a80d75e7632fc6058de645c91402dcc389"} Sep 30 14:14:16 crc kubenswrapper[4840]: I0930 14:14:16.005935 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf","Type":"ContainerDied","Data":"e289d0fde27d2ec83cbdf6dd13368f1d42bf91a3992499cf862f7683b6687fc3"} Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.025827 4840 generic.go:334] "Generic (PLEG): container finished" podID="4aa3849f-5f45-4c05-ac5a-dc8472d5acaf" containerID="cfcaf10f26382eada414852d9cab5083c54f9e5f2466f123779ce16711278cf9" exitCode=0 Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.025907 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf","Type":"ContainerDied","Data":"cfcaf10f26382eada414852d9cab5083c54f9e5f2466f123779ce16711278cf9"} Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.300604 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.300669 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.349729 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.367713 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.512364 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.512692 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.529067 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.556880 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.589681 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.626483 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-run-httpd\") pod \"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf\" (UID: \"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf\") " Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.626577 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-log-httpd\") pod \"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf\" (UID: \"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf\") " Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.626616 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-combined-ca-bundle\") pod \"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf\" (UID: \"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf\") " Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.626696 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-config-data\") pod \"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf\" (UID: \"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf\") " Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.626740 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9q7qd\" (UniqueName: \"kubernetes.io/projected/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-kube-api-access-9q7qd\") pod \"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf\" (UID: \"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf\") " Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.626826 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-sg-core-conf-yaml\") pod \"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf\" (UID: \"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf\") " Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.626862 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-scripts\") pod \"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf\" (UID: \"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf\") " Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.627109 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4aa3849f-5f45-4c05-ac5a-dc8472d5acaf" (UID: "4aa3849f-5f45-4c05-ac5a-dc8472d5acaf"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.627583 4840 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.627986 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4aa3849f-5f45-4c05-ac5a-dc8472d5acaf" (UID: "4aa3849f-5f45-4c05-ac5a-dc8472d5acaf"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.642978 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-kube-api-access-9q7qd" (OuterVolumeSpecName: "kube-api-access-9q7qd") pod "4aa3849f-5f45-4c05-ac5a-dc8472d5acaf" (UID: "4aa3849f-5f45-4c05-ac5a-dc8472d5acaf"). InnerVolumeSpecName "kube-api-access-9q7qd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.643134 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-scripts" (OuterVolumeSpecName: "scripts") pod "4aa3849f-5f45-4c05-ac5a-dc8472d5acaf" (UID: "4aa3849f-5f45-4c05-ac5a-dc8472d5acaf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.662967 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4aa3849f-5f45-4c05-ac5a-dc8472d5acaf" (UID: "4aa3849f-5f45-4c05-ac5a-dc8472d5acaf"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.719000 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-c6a4-account-create-r4fhv"] Sep 30 14:14:17 crc kubenswrapper[4840]: E0930 14:14:17.719657 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fe3cf3b-8d33-4689-86c9-3331fdcc6fac" containerName="mariadb-account-create" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.719684 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fe3cf3b-8d33-4689-86c9-3331fdcc6fac" containerName="mariadb-account-create" Sep 30 14:14:17 crc kubenswrapper[4840]: E0930 14:14:17.719706 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aa3849f-5f45-4c05-ac5a-dc8472d5acaf" containerName="sg-core" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.719714 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aa3849f-5f45-4c05-ac5a-dc8472d5acaf" containerName="sg-core" Sep 30 14:14:17 crc kubenswrapper[4840]: E0930 14:14:17.719727 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c2fd404-193d-4a27-92c9-395b2c0fcc02" containerName="mariadb-account-create" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.719737 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c2fd404-193d-4a27-92c9-395b2c0fcc02" containerName="mariadb-account-create" Sep 30 14:14:17 crc kubenswrapper[4840]: E0930 14:14:17.719752 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aa3849f-5f45-4c05-ac5a-dc8472d5acaf" containerName="ceilometer-central-agent" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.719760 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aa3849f-5f45-4c05-ac5a-dc8472d5acaf" containerName="ceilometer-central-agent" Sep 30 14:14:17 crc kubenswrapper[4840]: E0930 14:14:17.719782 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e9b87b2-4836-4857-8514-b2dcfe4de1f5" containerName="horizon-log" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.719792 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e9b87b2-4836-4857-8514-b2dcfe4de1f5" containerName="horizon-log" Sep 30 14:14:17 crc kubenswrapper[4840]: E0930 14:14:17.719807 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e9b87b2-4836-4857-8514-b2dcfe4de1f5" containerName="horizon" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.719814 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e9b87b2-4836-4857-8514-b2dcfe4de1f5" containerName="horizon" Sep 30 14:14:17 crc kubenswrapper[4840]: E0930 14:14:17.719872 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aa3849f-5f45-4c05-ac5a-dc8472d5acaf" containerName="proxy-httpd" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.719882 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aa3849f-5f45-4c05-ac5a-dc8472d5acaf" containerName="proxy-httpd" Sep 30 14:14:17 crc kubenswrapper[4840]: E0930 14:14:17.719906 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aa3849f-5f45-4c05-ac5a-dc8472d5acaf" containerName="ceilometer-notification-agent" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.719915 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aa3849f-5f45-4c05-ac5a-dc8472d5acaf" containerName="ceilometer-notification-agent" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.720136 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="4aa3849f-5f45-4c05-ac5a-dc8472d5acaf" containerName="ceilometer-notification-agent" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.720154 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c2fd404-193d-4a27-92c9-395b2c0fcc02" containerName="mariadb-account-create" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.720167 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="4aa3849f-5f45-4c05-ac5a-dc8472d5acaf" containerName="proxy-httpd" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.720181 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e9b87b2-4836-4857-8514-b2dcfe4de1f5" containerName="horizon-log" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.720203 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fe3cf3b-8d33-4689-86c9-3331fdcc6fac" containerName="mariadb-account-create" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.720213 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="4aa3849f-5f45-4c05-ac5a-dc8472d5acaf" containerName="ceilometer-central-agent" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.720227 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e9b87b2-4836-4857-8514-b2dcfe4de1f5" containerName="horizon" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.720242 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="4aa3849f-5f45-4c05-ac5a-dc8472d5acaf" containerName="sg-core" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.721043 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-c6a4-account-create-r4fhv" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.723921 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.732074 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9q7qd\" (UniqueName: \"kubernetes.io/projected/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-kube-api-access-9q7qd\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.732111 4840 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.732123 4840 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.732134 4840 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.736014 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-c6a4-account-create-r4fhv"] Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.740952 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4aa3849f-5f45-4c05-ac5a-dc8472d5acaf" (UID: "4aa3849f-5f45-4c05-ac5a-dc8472d5acaf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.765957 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-config-data" (OuterVolumeSpecName: "config-data") pod "4aa3849f-5f45-4c05-ac5a-dc8472d5acaf" (UID: "4aa3849f-5f45-4c05-ac5a-dc8472d5acaf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.834652 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zd4gx\" (UniqueName: \"kubernetes.io/projected/6f883b69-fee7-426a-9522-79e1c918e5fb-kube-api-access-zd4gx\") pod \"nova-cell0-c6a4-account-create-r4fhv\" (UID: \"6f883b69-fee7-426a-9522-79e1c918e5fb\") " pod="openstack/nova-cell0-c6a4-account-create-r4fhv" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.835461 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.835582 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.937165 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zd4gx\" (UniqueName: \"kubernetes.io/projected/6f883b69-fee7-426a-9522-79e1c918e5fb-kube-api-access-zd4gx\") pod \"nova-cell0-c6a4-account-create-r4fhv\" (UID: \"6f883b69-fee7-426a-9522-79e1c918e5fb\") " pod="openstack/nova-cell0-c6a4-account-create-r4fhv" Sep 30 14:14:17 crc kubenswrapper[4840]: I0930 14:14:17.952142 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zd4gx\" (UniqueName: \"kubernetes.io/projected/6f883b69-fee7-426a-9522-79e1c918e5fb-kube-api-access-zd4gx\") pod \"nova-cell0-c6a4-account-create-r4fhv\" (UID: \"6f883b69-fee7-426a-9522-79e1c918e5fb\") " pod="openstack/nova-cell0-c6a4-account-create-r4fhv" Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.040792 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4aa3849f-5f45-4c05-ac5a-dc8472d5acaf","Type":"ContainerDied","Data":"d05eee7cc0d5a444a422917ea1dd7c715fb505e754eee48067d533858389f278"} Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.040849 4840 scope.go:117] "RemoveContainer" containerID="63143e1a11e43c6259a3b43a048b0916e4859906600c2c4cc3f3b2db73fff7c4" Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.041029 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.041057 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.041083 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.042158 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.042201 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.055670 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-c6a4-account-create-r4fhv" Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.073351 4840 scope.go:117] "RemoveContainer" containerID="2108585a2ecf8b5b32dac6ce916c88a80d75e7632fc6058de645c91402dcc389" Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.084456 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.108522 4840 scope.go:117] "RemoveContainer" containerID="e289d0fde27d2ec83cbdf6dd13368f1d42bf91a3992499cf862f7683b6687fc3" Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.108659 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.177183 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4aa3849f-5f45-4c05-ac5a-dc8472d5acaf" path="/var/lib/kubelet/pods/4aa3849f-5f45-4c05-ac5a-dc8472d5acaf/volumes" Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.183730 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.192407 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.196505 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.196567 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.206142 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.224356 4840 scope.go:117] "RemoveContainer" containerID="cfcaf10f26382eada414852d9cab5083c54f9e5f2466f123779ce16711278cf9" Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.351384 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7\") " pod="openstack/ceilometer-0" Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.351829 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-scripts\") pod \"ceilometer-0\" (UID: \"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7\") " pod="openstack/ceilometer-0" Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.351864 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-run-httpd\") pod \"ceilometer-0\" (UID: \"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7\") " pod="openstack/ceilometer-0" Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.351905 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-config-data\") pod \"ceilometer-0\" (UID: \"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7\") " pod="openstack/ceilometer-0" Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.351961 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59b5t\" (UniqueName: \"kubernetes.io/projected/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-kube-api-access-59b5t\") pod \"ceilometer-0\" (UID: \"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7\") " pod="openstack/ceilometer-0" Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.352012 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7\") " pod="openstack/ceilometer-0" Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.352050 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-log-httpd\") pod \"ceilometer-0\" (UID: \"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7\") " pod="openstack/ceilometer-0" Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.453153 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-scripts\") pod \"ceilometer-0\" (UID: \"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7\") " pod="openstack/ceilometer-0" Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.453206 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-run-httpd\") pod \"ceilometer-0\" (UID: \"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7\") " pod="openstack/ceilometer-0" Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.453233 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-config-data\") pod \"ceilometer-0\" (UID: \"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7\") " pod="openstack/ceilometer-0" Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.453274 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59b5t\" (UniqueName: \"kubernetes.io/projected/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-kube-api-access-59b5t\") pod \"ceilometer-0\" (UID: \"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7\") " pod="openstack/ceilometer-0" Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.453334 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7\") " pod="openstack/ceilometer-0" Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.453387 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-log-httpd\") pod \"ceilometer-0\" (UID: \"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7\") " pod="openstack/ceilometer-0" Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.453443 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7\") " pod="openstack/ceilometer-0" Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.453702 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-run-httpd\") pod \"ceilometer-0\" (UID: \"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7\") " pod="openstack/ceilometer-0" Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.455128 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-log-httpd\") pod \"ceilometer-0\" (UID: \"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7\") " pod="openstack/ceilometer-0" Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.460423 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-scripts\") pod \"ceilometer-0\" (UID: \"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7\") " pod="openstack/ceilometer-0" Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.460541 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7\") " pod="openstack/ceilometer-0" Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.461480 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7\") " pod="openstack/ceilometer-0" Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.471225 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-config-data\") pod \"ceilometer-0\" (UID: \"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7\") " pod="openstack/ceilometer-0" Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.477080 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59b5t\" (UniqueName: \"kubernetes.io/projected/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-kube-api-access-59b5t\") pod \"ceilometer-0\" (UID: \"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7\") " pod="openstack/ceilometer-0" Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.525650 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:14:18 crc kubenswrapper[4840]: I0930 14:14:18.602489 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-c6a4-account-create-r4fhv"] Sep 30 14:14:19 crc kubenswrapper[4840]: W0930 14:14:19.024867 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbbea9705_7f8c_4b0f_9928_71c4b2fc7cc7.slice/crio-1cf9d6e6a5beedfdad77bf05c6021b276decf60d8e4708650b43900a086da61f WatchSource:0}: Error finding container 1cf9d6e6a5beedfdad77bf05c6021b276decf60d8e4708650b43900a086da61f: Status 404 returned error can't find the container with id 1cf9d6e6a5beedfdad77bf05c6021b276decf60d8e4708650b43900a086da61f Sep 30 14:14:19 crc kubenswrapper[4840]: I0930 14:14:19.025668 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:14:19 crc kubenswrapper[4840]: I0930 14:14:19.060810 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-c6a4-account-create-r4fhv" event={"ID":"6f883b69-fee7-426a-9522-79e1c918e5fb","Type":"ContainerStarted","Data":"455ebb76c7cfb7ea9121ca949e9eae745536c621e9b7c76bc7e8ad9ec7eea647"} Sep 30 14:14:19 crc kubenswrapper[4840]: I0930 14:14:19.060879 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-c6a4-account-create-r4fhv" event={"ID":"6f883b69-fee7-426a-9522-79e1c918e5fb","Type":"ContainerStarted","Data":"cad39451ea2c8743f8d23194400f2d9833697559678ba494c08a8b730650d5bc"} Sep 30 14:14:19 crc kubenswrapper[4840]: I0930 14:14:19.062631 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7","Type":"ContainerStarted","Data":"1cf9d6e6a5beedfdad77bf05c6021b276decf60d8e4708650b43900a086da61f"} Sep 30 14:14:20 crc kubenswrapper[4840]: I0930 14:14:20.083414 4840 generic.go:334] "Generic (PLEG): container finished" podID="6f883b69-fee7-426a-9522-79e1c918e5fb" containerID="455ebb76c7cfb7ea9121ca949e9eae745536c621e9b7c76bc7e8ad9ec7eea647" exitCode=0 Sep 30 14:14:20 crc kubenswrapper[4840]: I0930 14:14:20.084034 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-c6a4-account-create-r4fhv" event={"ID":"6f883b69-fee7-426a-9522-79e1c918e5fb","Type":"ContainerDied","Data":"455ebb76c7cfb7ea9121ca949e9eae745536c621e9b7c76bc7e8ad9ec7eea647"} Sep 30 14:14:20 crc kubenswrapper[4840]: I0930 14:14:20.088420 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Sep 30 14:14:20 crc kubenswrapper[4840]: I0930 14:14:20.088505 4840 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 14:14:20 crc kubenswrapper[4840]: I0930 14:14:20.194254 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Sep 30 14:14:20 crc kubenswrapper[4840]: I0930 14:14:20.323918 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Sep 30 14:14:20 crc kubenswrapper[4840]: I0930 14:14:20.324218 4840 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 14:14:20 crc kubenswrapper[4840]: I0930 14:14:20.327935 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Sep 30 14:14:20 crc kubenswrapper[4840]: I0930 14:14:20.473638 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-c6a4-account-create-r4fhv" Sep 30 14:14:20 crc kubenswrapper[4840]: I0930 14:14:20.605157 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zd4gx\" (UniqueName: \"kubernetes.io/projected/6f883b69-fee7-426a-9522-79e1c918e5fb-kube-api-access-zd4gx\") pod \"6f883b69-fee7-426a-9522-79e1c918e5fb\" (UID: \"6f883b69-fee7-426a-9522-79e1c918e5fb\") " Sep 30 14:14:20 crc kubenswrapper[4840]: I0930 14:14:20.615764 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f883b69-fee7-426a-9522-79e1c918e5fb-kube-api-access-zd4gx" (OuterVolumeSpecName: "kube-api-access-zd4gx") pod "6f883b69-fee7-426a-9522-79e1c918e5fb" (UID: "6f883b69-fee7-426a-9522-79e1c918e5fb"). InnerVolumeSpecName "kube-api-access-zd4gx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:14:20 crc kubenswrapper[4840]: I0930 14:14:20.707641 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zd4gx\" (UniqueName: \"kubernetes.io/projected/6f883b69-fee7-426a-9522-79e1c918e5fb-kube-api-access-zd4gx\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:21 crc kubenswrapper[4840]: I0930 14:14:21.094335 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-c6a4-account-create-r4fhv" Sep 30 14:14:21 crc kubenswrapper[4840]: I0930 14:14:21.094328 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-c6a4-account-create-r4fhv" event={"ID":"6f883b69-fee7-426a-9522-79e1c918e5fb","Type":"ContainerDied","Data":"cad39451ea2c8743f8d23194400f2d9833697559678ba494c08a8b730650d5bc"} Sep 30 14:14:21 crc kubenswrapper[4840]: I0930 14:14:21.095609 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cad39451ea2c8743f8d23194400f2d9833697559678ba494c08a8b730650d5bc" Sep 30 14:14:21 crc kubenswrapper[4840]: I0930 14:14:21.097034 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7","Type":"ContainerStarted","Data":"f0d714205f0ed2f357e46f0e6fa8857ff6cc3fcd1546a97c565e5af0b54e026a"} Sep 30 14:14:21 crc kubenswrapper[4840]: I0930 14:14:21.097075 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7","Type":"ContainerStarted","Data":"59089db8291df07aace5e8394e31eaf5bbb156b34e4ea37b0b4a1cfa7e3c1767"} Sep 30 14:14:22 crc kubenswrapper[4840]: I0930 14:14:22.130491 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7","Type":"ContainerStarted","Data":"41d28c86528720a96046e6f29345bba6714f32b13934518dafda2f4e97b0781d"} Sep 30 14:14:22 crc kubenswrapper[4840]: I0930 14:14:22.898969 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-nlp8v"] Sep 30 14:14:22 crc kubenswrapper[4840]: E0930 14:14:22.899704 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f883b69-fee7-426a-9522-79e1c918e5fb" containerName="mariadb-account-create" Sep 30 14:14:22 crc kubenswrapper[4840]: I0930 14:14:22.899797 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f883b69-fee7-426a-9522-79e1c918e5fb" containerName="mariadb-account-create" Sep 30 14:14:22 crc kubenswrapper[4840]: I0930 14:14:22.900068 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f883b69-fee7-426a-9522-79e1c918e5fb" containerName="mariadb-account-create" Sep 30 14:14:22 crc kubenswrapper[4840]: I0930 14:14:22.900743 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-nlp8v" Sep 30 14:14:22 crc kubenswrapper[4840]: I0930 14:14:22.903635 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Sep 30 14:14:22 crc kubenswrapper[4840]: I0930 14:14:22.903689 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Sep 30 14:14:22 crc kubenswrapper[4840]: I0930 14:14:22.903776 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-bgvhh" Sep 30 14:14:22 crc kubenswrapper[4840]: I0930 14:14:22.913004 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-nlp8v"] Sep 30 14:14:23 crc kubenswrapper[4840]: I0930 14:14:23.047458 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7e57e20-fd49-4406-9814-1486cc9efdf0-config-data\") pod \"nova-cell0-conductor-db-sync-nlp8v\" (UID: \"d7e57e20-fd49-4406-9814-1486cc9efdf0\") " pod="openstack/nova-cell0-conductor-db-sync-nlp8v" Sep 30 14:14:23 crc kubenswrapper[4840]: I0930 14:14:23.047785 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7e57e20-fd49-4406-9814-1486cc9efdf0-scripts\") pod \"nova-cell0-conductor-db-sync-nlp8v\" (UID: \"d7e57e20-fd49-4406-9814-1486cc9efdf0\") " pod="openstack/nova-cell0-conductor-db-sync-nlp8v" Sep 30 14:14:23 crc kubenswrapper[4840]: I0930 14:14:23.047892 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhqw8\" (UniqueName: \"kubernetes.io/projected/d7e57e20-fd49-4406-9814-1486cc9efdf0-kube-api-access-jhqw8\") pod \"nova-cell0-conductor-db-sync-nlp8v\" (UID: \"d7e57e20-fd49-4406-9814-1486cc9efdf0\") " pod="openstack/nova-cell0-conductor-db-sync-nlp8v" Sep 30 14:14:23 crc kubenswrapper[4840]: I0930 14:14:23.048037 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7e57e20-fd49-4406-9814-1486cc9efdf0-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-nlp8v\" (UID: \"d7e57e20-fd49-4406-9814-1486cc9efdf0\") " pod="openstack/nova-cell0-conductor-db-sync-nlp8v" Sep 30 14:14:23 crc kubenswrapper[4840]: I0930 14:14:23.149451 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7e57e20-fd49-4406-9814-1486cc9efdf0-scripts\") pod \"nova-cell0-conductor-db-sync-nlp8v\" (UID: \"d7e57e20-fd49-4406-9814-1486cc9efdf0\") " pod="openstack/nova-cell0-conductor-db-sync-nlp8v" Sep 30 14:14:23 crc kubenswrapper[4840]: I0930 14:14:23.149717 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhqw8\" (UniqueName: \"kubernetes.io/projected/d7e57e20-fd49-4406-9814-1486cc9efdf0-kube-api-access-jhqw8\") pod \"nova-cell0-conductor-db-sync-nlp8v\" (UID: \"d7e57e20-fd49-4406-9814-1486cc9efdf0\") " pod="openstack/nova-cell0-conductor-db-sync-nlp8v" Sep 30 14:14:23 crc kubenswrapper[4840]: I0930 14:14:23.149819 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7e57e20-fd49-4406-9814-1486cc9efdf0-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-nlp8v\" (UID: \"d7e57e20-fd49-4406-9814-1486cc9efdf0\") " pod="openstack/nova-cell0-conductor-db-sync-nlp8v" Sep 30 14:14:23 crc kubenswrapper[4840]: I0930 14:14:23.149863 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7e57e20-fd49-4406-9814-1486cc9efdf0-config-data\") pod \"nova-cell0-conductor-db-sync-nlp8v\" (UID: \"d7e57e20-fd49-4406-9814-1486cc9efdf0\") " pod="openstack/nova-cell0-conductor-db-sync-nlp8v" Sep 30 14:14:23 crc kubenswrapper[4840]: I0930 14:14:23.153801 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7e57e20-fd49-4406-9814-1486cc9efdf0-scripts\") pod \"nova-cell0-conductor-db-sync-nlp8v\" (UID: \"d7e57e20-fd49-4406-9814-1486cc9efdf0\") " pod="openstack/nova-cell0-conductor-db-sync-nlp8v" Sep 30 14:14:23 crc kubenswrapper[4840]: I0930 14:14:23.154079 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7e57e20-fd49-4406-9814-1486cc9efdf0-config-data\") pod \"nova-cell0-conductor-db-sync-nlp8v\" (UID: \"d7e57e20-fd49-4406-9814-1486cc9efdf0\") " pod="openstack/nova-cell0-conductor-db-sync-nlp8v" Sep 30 14:14:23 crc kubenswrapper[4840]: I0930 14:14:23.156168 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7e57e20-fd49-4406-9814-1486cc9efdf0-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-nlp8v\" (UID: \"d7e57e20-fd49-4406-9814-1486cc9efdf0\") " pod="openstack/nova-cell0-conductor-db-sync-nlp8v" Sep 30 14:14:23 crc kubenswrapper[4840]: I0930 14:14:23.169090 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhqw8\" (UniqueName: \"kubernetes.io/projected/d7e57e20-fd49-4406-9814-1486cc9efdf0-kube-api-access-jhqw8\") pod \"nova-cell0-conductor-db-sync-nlp8v\" (UID: \"d7e57e20-fd49-4406-9814-1486cc9efdf0\") " pod="openstack/nova-cell0-conductor-db-sync-nlp8v" Sep 30 14:14:23 crc kubenswrapper[4840]: I0930 14:14:23.219921 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-nlp8v" Sep 30 14:14:23 crc kubenswrapper[4840]: I0930 14:14:23.713127 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-nlp8v"] Sep 30 14:14:23 crc kubenswrapper[4840]: W0930 14:14:23.719719 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7e57e20_fd49_4406_9814_1486cc9efdf0.slice/crio-77237d21476f98c7971b9237c78482ec28734e1382088face1e615691cd26659 WatchSource:0}: Error finding container 77237d21476f98c7971b9237c78482ec28734e1382088face1e615691cd26659: Status 404 returned error can't find the container with id 77237d21476f98c7971b9237c78482ec28734e1382088face1e615691cd26659 Sep 30 14:14:24 crc kubenswrapper[4840]: I0930 14:14:24.153155 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-nlp8v" event={"ID":"d7e57e20-fd49-4406-9814-1486cc9efdf0","Type":"ContainerStarted","Data":"77237d21476f98c7971b9237c78482ec28734e1382088face1e615691cd26659"} Sep 30 14:14:24 crc kubenswrapper[4840]: I0930 14:14:24.156751 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7","Type":"ContainerStarted","Data":"99cbad6f57500a2b93ad3084b0a1c5399fc9055fbb48c1619bf86d6226c18c0c"} Sep 30 14:14:24 crc kubenswrapper[4840]: I0930 14:14:24.157036 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 30 14:14:24 crc kubenswrapper[4840]: I0930 14:14:24.187059 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.153466764 podStartE2EDuration="6.187035721s" podCreationTimestamp="2025-09-30 14:14:18 +0000 UTC" firstStartedPulling="2025-09-30 14:14:19.027441949 +0000 UTC m=+1087.656528372" lastFinishedPulling="2025-09-30 14:14:23.061010906 +0000 UTC m=+1091.690097329" observedRunningTime="2025-09-30 14:14:24.181035694 +0000 UTC m=+1092.810122127" watchObservedRunningTime="2025-09-30 14:14:24.187035721 +0000 UTC m=+1092.816122144" Sep 30 14:14:26 crc kubenswrapper[4840]: I0930 14:14:26.895422 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:14:26 crc kubenswrapper[4840]: I0930 14:14:26.896016 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7" containerName="ceilometer-central-agent" containerID="cri-o://59089db8291df07aace5e8394e31eaf5bbb156b34e4ea37b0b4a1cfa7e3c1767" gracePeriod=30 Sep 30 14:14:26 crc kubenswrapper[4840]: I0930 14:14:26.896416 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7" containerName="proxy-httpd" containerID="cri-o://99cbad6f57500a2b93ad3084b0a1c5399fc9055fbb48c1619bf86d6226c18c0c" gracePeriod=30 Sep 30 14:14:26 crc kubenswrapper[4840]: I0930 14:14:26.896474 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7" containerName="sg-core" containerID="cri-o://41d28c86528720a96046e6f29345bba6714f32b13934518dafda2f4e97b0781d" gracePeriod=30 Sep 30 14:14:26 crc kubenswrapper[4840]: I0930 14:14:26.896519 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7" containerName="ceilometer-notification-agent" containerID="cri-o://f0d714205f0ed2f357e46f0e6fa8857ff6cc3fcd1546a97c565e5af0b54e026a" gracePeriod=30 Sep 30 14:14:27 crc kubenswrapper[4840]: I0930 14:14:27.189477 4840 generic.go:334] "Generic (PLEG): container finished" podID="bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7" containerID="99cbad6f57500a2b93ad3084b0a1c5399fc9055fbb48c1619bf86d6226c18c0c" exitCode=0 Sep 30 14:14:27 crc kubenswrapper[4840]: I0930 14:14:27.189757 4840 generic.go:334] "Generic (PLEG): container finished" podID="bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7" containerID="41d28c86528720a96046e6f29345bba6714f32b13934518dafda2f4e97b0781d" exitCode=2 Sep 30 14:14:27 crc kubenswrapper[4840]: I0930 14:14:27.189572 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7","Type":"ContainerDied","Data":"99cbad6f57500a2b93ad3084b0a1c5399fc9055fbb48c1619bf86d6226c18c0c"} Sep 30 14:14:27 crc kubenswrapper[4840]: I0930 14:14:27.189789 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7","Type":"ContainerDied","Data":"41d28c86528720a96046e6f29345bba6714f32b13934518dafda2f4e97b0781d"} Sep 30 14:14:28 crc kubenswrapper[4840]: I0930 14:14:28.200981 4840 generic.go:334] "Generic (PLEG): container finished" podID="bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7" containerID="f0d714205f0ed2f357e46f0e6fa8857ff6cc3fcd1546a97c565e5af0b54e026a" exitCode=0 Sep 30 14:14:28 crc kubenswrapper[4840]: I0930 14:14:28.201026 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7","Type":"ContainerDied","Data":"f0d714205f0ed2f357e46f0e6fa8857ff6cc3fcd1546a97c565e5af0b54e026a"} Sep 30 14:14:31 crc kubenswrapper[4840]: I0930 14:14:31.277621 4840 generic.go:334] "Generic (PLEG): container finished" podID="bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7" containerID="59089db8291df07aace5e8394e31eaf5bbb156b34e4ea37b0b4a1cfa7e3c1767" exitCode=0 Sep 30 14:14:31 crc kubenswrapper[4840]: I0930 14:14:31.278018 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7","Type":"ContainerDied","Data":"59089db8291df07aace5e8394e31eaf5bbb156b34e4ea37b0b4a1cfa7e3c1767"} Sep 30 14:14:32 crc kubenswrapper[4840]: I0930 14:14:32.439155 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:14:32 crc kubenswrapper[4840]: I0930 14:14:32.539526 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-run-httpd\") pod \"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7\" (UID: \"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7\") " Sep 30 14:14:32 crc kubenswrapper[4840]: I0930 14:14:32.539587 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-scripts\") pod \"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7\" (UID: \"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7\") " Sep 30 14:14:32 crc kubenswrapper[4840]: I0930 14:14:32.539617 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-combined-ca-bundle\") pod \"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7\" (UID: \"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7\") " Sep 30 14:14:32 crc kubenswrapper[4840]: I0930 14:14:32.539691 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-config-data\") pod \"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7\" (UID: \"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7\") " Sep 30 14:14:32 crc kubenswrapper[4840]: I0930 14:14:32.539750 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-sg-core-conf-yaml\") pod \"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7\" (UID: \"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7\") " Sep 30 14:14:32 crc kubenswrapper[4840]: I0930 14:14:32.539826 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-log-httpd\") pod \"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7\" (UID: \"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7\") " Sep 30 14:14:32 crc kubenswrapper[4840]: I0930 14:14:32.539857 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59b5t\" (UniqueName: \"kubernetes.io/projected/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-kube-api-access-59b5t\") pod \"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7\" (UID: \"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7\") " Sep 30 14:14:32 crc kubenswrapper[4840]: I0930 14:14:32.540073 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7" (UID: "bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:14:32 crc kubenswrapper[4840]: I0930 14:14:32.540644 4840 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:32 crc kubenswrapper[4840]: I0930 14:14:32.541189 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7" (UID: "bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:14:32 crc kubenswrapper[4840]: I0930 14:14:32.544501 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-kube-api-access-59b5t" (OuterVolumeSpecName: "kube-api-access-59b5t") pod "bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7" (UID: "bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7"). InnerVolumeSpecName "kube-api-access-59b5t". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:14:32 crc kubenswrapper[4840]: I0930 14:14:32.544533 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-scripts" (OuterVolumeSpecName: "scripts") pod "bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7" (UID: "bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:14:32 crc kubenswrapper[4840]: I0930 14:14:32.567675 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7" (UID: "bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:14:32 crc kubenswrapper[4840]: I0930 14:14:32.605527 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7" (UID: "bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:14:32 crc kubenswrapper[4840]: I0930 14:14:32.628073 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-config-data" (OuterVolumeSpecName: "config-data") pod "bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7" (UID: "bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:14:32 crc kubenswrapper[4840]: I0930 14:14:32.642856 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:32 crc kubenswrapper[4840]: I0930 14:14:32.642899 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:32 crc kubenswrapper[4840]: I0930 14:14:32.642918 4840 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:32 crc kubenswrapper[4840]: I0930 14:14:32.642936 4840 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:32 crc kubenswrapper[4840]: I0930 14:14:32.642953 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59b5t\" (UniqueName: \"kubernetes.io/projected/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-kube-api-access-59b5t\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:32 crc kubenswrapper[4840]: I0930 14:14:32.642972 4840 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.311820 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-nlp8v" event={"ID":"d7e57e20-fd49-4406-9814-1486cc9efdf0","Type":"ContainerStarted","Data":"ea758c4f9027095d42744a11ef0b508192d6c55fb24d66b7de02b3df32145dcf"} Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.314793 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7","Type":"ContainerDied","Data":"1cf9d6e6a5beedfdad77bf05c6021b276decf60d8e4708650b43900a086da61f"} Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.314839 4840 scope.go:117] "RemoveContainer" containerID="99cbad6f57500a2b93ad3084b0a1c5399fc9055fbb48c1619bf86d6226c18c0c" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.314983 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.348742 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-nlp8v" podStartSLOduration=2.829963702 podStartE2EDuration="11.348725654s" podCreationTimestamp="2025-09-30 14:14:22 +0000 UTC" firstStartedPulling="2025-09-30 14:14:23.722764853 +0000 UTC m=+1092.351851276" lastFinishedPulling="2025-09-30 14:14:32.241526805 +0000 UTC m=+1100.870613228" observedRunningTime="2025-09-30 14:14:33.33899956 +0000 UTC m=+1101.968085983" watchObservedRunningTime="2025-09-30 14:14:33.348725654 +0000 UTC m=+1101.977812077" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.371089 4840 scope.go:117] "RemoveContainer" containerID="41d28c86528720a96046e6f29345bba6714f32b13934518dafda2f4e97b0781d" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.373033 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.385122 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.401338 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:14:33 crc kubenswrapper[4840]: E0930 14:14:33.401747 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7" containerName="ceilometer-central-agent" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.401764 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7" containerName="ceilometer-central-agent" Sep 30 14:14:33 crc kubenswrapper[4840]: E0930 14:14:33.401791 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7" containerName="proxy-httpd" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.401799 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7" containerName="proxy-httpd" Sep 30 14:14:33 crc kubenswrapper[4840]: E0930 14:14:33.401807 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7" containerName="sg-core" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.401813 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7" containerName="sg-core" Sep 30 14:14:33 crc kubenswrapper[4840]: E0930 14:14:33.401826 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7" containerName="ceilometer-notification-agent" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.401832 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7" containerName="ceilometer-notification-agent" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.402000 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7" containerName="ceilometer-notification-agent" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.402007 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7" containerName="proxy-httpd" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.402020 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7" containerName="ceilometer-central-agent" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.402030 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7" containerName="sg-core" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.403576 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.406672 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.406915 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.414739 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.421622 4840 scope.go:117] "RemoveContainer" containerID="f0d714205f0ed2f357e46f0e6fa8857ff6cc3fcd1546a97c565e5af0b54e026a" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.448390 4840 scope.go:117] "RemoveContainer" containerID="59089db8291df07aace5e8394e31eaf5bbb156b34e4ea37b0b4a1cfa7e3c1767" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.457339 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/398ac086-cb37-46db-a4fa-e964bf468b81-config-data\") pod \"ceilometer-0\" (UID: \"398ac086-cb37-46db-a4fa-e964bf468b81\") " pod="openstack/ceilometer-0" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.457385 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2h7xh\" (UniqueName: \"kubernetes.io/projected/398ac086-cb37-46db-a4fa-e964bf468b81-kube-api-access-2h7xh\") pod \"ceilometer-0\" (UID: \"398ac086-cb37-46db-a4fa-e964bf468b81\") " pod="openstack/ceilometer-0" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.457411 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/398ac086-cb37-46db-a4fa-e964bf468b81-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"398ac086-cb37-46db-a4fa-e964bf468b81\") " pod="openstack/ceilometer-0" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.457468 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/398ac086-cb37-46db-a4fa-e964bf468b81-log-httpd\") pod \"ceilometer-0\" (UID: \"398ac086-cb37-46db-a4fa-e964bf468b81\") " pod="openstack/ceilometer-0" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.457490 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/398ac086-cb37-46db-a4fa-e964bf468b81-run-httpd\") pod \"ceilometer-0\" (UID: \"398ac086-cb37-46db-a4fa-e964bf468b81\") " pod="openstack/ceilometer-0" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.457541 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/398ac086-cb37-46db-a4fa-e964bf468b81-scripts\") pod \"ceilometer-0\" (UID: \"398ac086-cb37-46db-a4fa-e964bf468b81\") " pod="openstack/ceilometer-0" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.457601 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/398ac086-cb37-46db-a4fa-e964bf468b81-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"398ac086-cb37-46db-a4fa-e964bf468b81\") " pod="openstack/ceilometer-0" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.559382 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/398ac086-cb37-46db-a4fa-e964bf468b81-log-httpd\") pod \"ceilometer-0\" (UID: \"398ac086-cb37-46db-a4fa-e964bf468b81\") " pod="openstack/ceilometer-0" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.559662 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/398ac086-cb37-46db-a4fa-e964bf468b81-run-httpd\") pod \"ceilometer-0\" (UID: \"398ac086-cb37-46db-a4fa-e964bf468b81\") " pod="openstack/ceilometer-0" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.559861 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/398ac086-cb37-46db-a4fa-e964bf468b81-scripts\") pod \"ceilometer-0\" (UID: \"398ac086-cb37-46db-a4fa-e964bf468b81\") " pod="openstack/ceilometer-0" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.559963 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/398ac086-cb37-46db-a4fa-e964bf468b81-log-httpd\") pod \"ceilometer-0\" (UID: \"398ac086-cb37-46db-a4fa-e964bf468b81\") " pod="openstack/ceilometer-0" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.560061 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/398ac086-cb37-46db-a4fa-e964bf468b81-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"398ac086-cb37-46db-a4fa-e964bf468b81\") " pod="openstack/ceilometer-0" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.560212 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/398ac086-cb37-46db-a4fa-e964bf468b81-config-data\") pod \"ceilometer-0\" (UID: \"398ac086-cb37-46db-a4fa-e964bf468b81\") " pod="openstack/ceilometer-0" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.560313 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2h7xh\" (UniqueName: \"kubernetes.io/projected/398ac086-cb37-46db-a4fa-e964bf468b81-kube-api-access-2h7xh\") pod \"ceilometer-0\" (UID: \"398ac086-cb37-46db-a4fa-e964bf468b81\") " pod="openstack/ceilometer-0" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.560424 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/398ac086-cb37-46db-a4fa-e964bf468b81-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"398ac086-cb37-46db-a4fa-e964bf468b81\") " pod="openstack/ceilometer-0" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.560076 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/398ac086-cb37-46db-a4fa-e964bf468b81-run-httpd\") pod \"ceilometer-0\" (UID: \"398ac086-cb37-46db-a4fa-e964bf468b81\") " pod="openstack/ceilometer-0" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.564508 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/398ac086-cb37-46db-a4fa-e964bf468b81-scripts\") pod \"ceilometer-0\" (UID: \"398ac086-cb37-46db-a4fa-e964bf468b81\") " pod="openstack/ceilometer-0" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.566541 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/398ac086-cb37-46db-a4fa-e964bf468b81-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"398ac086-cb37-46db-a4fa-e964bf468b81\") " pod="openstack/ceilometer-0" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.567031 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/398ac086-cb37-46db-a4fa-e964bf468b81-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"398ac086-cb37-46db-a4fa-e964bf468b81\") " pod="openstack/ceilometer-0" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.572267 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/398ac086-cb37-46db-a4fa-e964bf468b81-config-data\") pod \"ceilometer-0\" (UID: \"398ac086-cb37-46db-a4fa-e964bf468b81\") " pod="openstack/ceilometer-0" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.577813 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2h7xh\" (UniqueName: \"kubernetes.io/projected/398ac086-cb37-46db-a4fa-e964bf468b81-kube-api-access-2h7xh\") pod \"ceilometer-0\" (UID: \"398ac086-cb37-46db-a4fa-e964bf468b81\") " pod="openstack/ceilometer-0" Sep 30 14:14:33 crc kubenswrapper[4840]: I0930 14:14:33.731456 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:14:34 crc kubenswrapper[4840]: I0930 14:14:34.126605 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7" path="/var/lib/kubelet/pods/bbea9705-7f8c-4b0f-9928-71c4b2fc7cc7/volumes" Sep 30 14:14:34 crc kubenswrapper[4840]: I0930 14:14:34.187921 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:14:34 crc kubenswrapper[4840]: I0930 14:14:34.324279 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"398ac086-cb37-46db-a4fa-e964bf468b81","Type":"ContainerStarted","Data":"1e0b25c5e709e9e43348b25c191d46dbc539dc9902622005c297e02c8e11a8d2"} Sep 30 14:14:35 crc kubenswrapper[4840]: I0930 14:14:35.347456 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"398ac086-cb37-46db-a4fa-e964bf468b81","Type":"ContainerStarted","Data":"abf8546fd8c0c1e84bbe02106ed1287e7966858e265c9a54580b44162ded1c2f"} Sep 30 14:14:35 crc kubenswrapper[4840]: I0930 14:14:35.367076 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:14:36 crc kubenswrapper[4840]: I0930 14:14:36.360487 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"398ac086-cb37-46db-a4fa-e964bf468b81","Type":"ContainerStarted","Data":"658f57946407b68ab2e09b557a0c2341838801cfbfb57c5cd469301197d45699"} Sep 30 14:14:38 crc kubenswrapper[4840]: I0930 14:14:38.391722 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"398ac086-cb37-46db-a4fa-e964bf468b81","Type":"ContainerStarted","Data":"274d7e35969926c142e97bdd427d59044bf48e96015d9626c7a3b9a3b23ee7a9"} Sep 30 14:14:39 crc kubenswrapper[4840]: I0930 14:14:39.404747 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"398ac086-cb37-46db-a4fa-e964bf468b81","Type":"ContainerStarted","Data":"4cc932978f9822386e3a5f8f4135308469aed6b1b0a4f69f8c362a02759d7ab6"} Sep 30 14:14:39 crc kubenswrapper[4840]: I0930 14:14:39.405039 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="398ac086-cb37-46db-a4fa-e964bf468b81" containerName="ceilometer-notification-agent" containerID="cri-o://658f57946407b68ab2e09b557a0c2341838801cfbfb57c5cd469301197d45699" gracePeriod=30 Sep 30 14:14:39 crc kubenswrapper[4840]: I0930 14:14:39.405083 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="398ac086-cb37-46db-a4fa-e964bf468b81" containerName="sg-core" containerID="cri-o://274d7e35969926c142e97bdd427d59044bf48e96015d9626c7a3b9a3b23ee7a9" gracePeriod=30 Sep 30 14:14:39 crc kubenswrapper[4840]: I0930 14:14:39.404929 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="398ac086-cb37-46db-a4fa-e964bf468b81" containerName="ceilometer-central-agent" containerID="cri-o://abf8546fd8c0c1e84bbe02106ed1287e7966858e265c9a54580b44162ded1c2f" gracePeriod=30 Sep 30 14:14:39 crc kubenswrapper[4840]: I0930 14:14:39.404981 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="398ac086-cb37-46db-a4fa-e964bf468b81" containerName="proxy-httpd" containerID="cri-o://4cc932978f9822386e3a5f8f4135308469aed6b1b0a4f69f8c362a02759d7ab6" gracePeriod=30 Sep 30 14:14:39 crc kubenswrapper[4840]: I0930 14:14:39.405258 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 30 14:14:39 crc kubenswrapper[4840]: I0930 14:14:39.433927 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.522841036 podStartE2EDuration="6.433902877s" podCreationTimestamp="2025-09-30 14:14:33 +0000 UTC" firstStartedPulling="2025-09-30 14:14:34.194998491 +0000 UTC m=+1102.824084914" lastFinishedPulling="2025-09-30 14:14:39.106060332 +0000 UTC m=+1107.735146755" observedRunningTime="2025-09-30 14:14:39.425348884 +0000 UTC m=+1108.054435307" watchObservedRunningTime="2025-09-30 14:14:39.433902877 +0000 UTC m=+1108.062989300" Sep 30 14:14:40 crc kubenswrapper[4840]: I0930 14:14:40.417450 4840 generic.go:334] "Generic (PLEG): container finished" podID="398ac086-cb37-46db-a4fa-e964bf468b81" containerID="4cc932978f9822386e3a5f8f4135308469aed6b1b0a4f69f8c362a02759d7ab6" exitCode=0 Sep 30 14:14:40 crc kubenswrapper[4840]: I0930 14:14:40.417924 4840 generic.go:334] "Generic (PLEG): container finished" podID="398ac086-cb37-46db-a4fa-e964bf468b81" containerID="274d7e35969926c142e97bdd427d59044bf48e96015d9626c7a3b9a3b23ee7a9" exitCode=2 Sep 30 14:14:40 crc kubenswrapper[4840]: I0930 14:14:40.417937 4840 generic.go:334] "Generic (PLEG): container finished" podID="398ac086-cb37-46db-a4fa-e964bf468b81" containerID="658f57946407b68ab2e09b557a0c2341838801cfbfb57c5cd469301197d45699" exitCode=0 Sep 30 14:14:40 crc kubenswrapper[4840]: I0930 14:14:40.417511 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"398ac086-cb37-46db-a4fa-e964bf468b81","Type":"ContainerDied","Data":"4cc932978f9822386e3a5f8f4135308469aed6b1b0a4f69f8c362a02759d7ab6"} Sep 30 14:14:40 crc kubenswrapper[4840]: I0930 14:14:40.417979 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"398ac086-cb37-46db-a4fa-e964bf468b81","Type":"ContainerDied","Data":"274d7e35969926c142e97bdd427d59044bf48e96015d9626c7a3b9a3b23ee7a9"} Sep 30 14:14:40 crc kubenswrapper[4840]: I0930 14:14:40.417998 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"398ac086-cb37-46db-a4fa-e964bf468b81","Type":"ContainerDied","Data":"658f57946407b68ab2e09b557a0c2341838801cfbfb57c5cd469301197d45699"} Sep 30 14:14:41 crc kubenswrapper[4840]: I0930 14:14:41.431091 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"398ac086-cb37-46db-a4fa-e964bf468b81","Type":"ContainerDied","Data":"abf8546fd8c0c1e84bbe02106ed1287e7966858e265c9a54580b44162ded1c2f"} Sep 30 14:14:41 crc kubenswrapper[4840]: I0930 14:14:41.431029 4840 generic.go:334] "Generic (PLEG): container finished" podID="398ac086-cb37-46db-a4fa-e964bf468b81" containerID="abf8546fd8c0c1e84bbe02106ed1287e7966858e265c9a54580b44162ded1c2f" exitCode=0 Sep 30 14:14:41 crc kubenswrapper[4840]: I0930 14:14:41.431597 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"398ac086-cb37-46db-a4fa-e964bf468b81","Type":"ContainerDied","Data":"1e0b25c5e709e9e43348b25c191d46dbc539dc9902622005c297e02c8e11a8d2"} Sep 30 14:14:41 crc kubenswrapper[4840]: I0930 14:14:41.431633 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e0b25c5e709e9e43348b25c191d46dbc539dc9902622005c297e02c8e11a8d2" Sep 30 14:14:41 crc kubenswrapper[4840]: I0930 14:14:41.474834 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:14:41 crc kubenswrapper[4840]: I0930 14:14:41.620382 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/398ac086-cb37-46db-a4fa-e964bf468b81-sg-core-conf-yaml\") pod \"398ac086-cb37-46db-a4fa-e964bf468b81\" (UID: \"398ac086-cb37-46db-a4fa-e964bf468b81\") " Sep 30 14:14:41 crc kubenswrapper[4840]: I0930 14:14:41.620442 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/398ac086-cb37-46db-a4fa-e964bf468b81-config-data\") pod \"398ac086-cb37-46db-a4fa-e964bf468b81\" (UID: \"398ac086-cb37-46db-a4fa-e964bf468b81\") " Sep 30 14:14:41 crc kubenswrapper[4840]: I0930 14:14:41.620510 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/398ac086-cb37-46db-a4fa-e964bf468b81-combined-ca-bundle\") pod \"398ac086-cb37-46db-a4fa-e964bf468b81\" (UID: \"398ac086-cb37-46db-a4fa-e964bf468b81\") " Sep 30 14:14:41 crc kubenswrapper[4840]: I0930 14:14:41.620630 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/398ac086-cb37-46db-a4fa-e964bf468b81-scripts\") pod \"398ac086-cb37-46db-a4fa-e964bf468b81\" (UID: \"398ac086-cb37-46db-a4fa-e964bf468b81\") " Sep 30 14:14:41 crc kubenswrapper[4840]: I0930 14:14:41.620688 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/398ac086-cb37-46db-a4fa-e964bf468b81-run-httpd\") pod \"398ac086-cb37-46db-a4fa-e964bf468b81\" (UID: \"398ac086-cb37-46db-a4fa-e964bf468b81\") " Sep 30 14:14:41 crc kubenswrapper[4840]: I0930 14:14:41.620729 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2h7xh\" (UniqueName: \"kubernetes.io/projected/398ac086-cb37-46db-a4fa-e964bf468b81-kube-api-access-2h7xh\") pod \"398ac086-cb37-46db-a4fa-e964bf468b81\" (UID: \"398ac086-cb37-46db-a4fa-e964bf468b81\") " Sep 30 14:14:41 crc kubenswrapper[4840]: I0930 14:14:41.620969 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/398ac086-cb37-46db-a4fa-e964bf468b81-log-httpd\") pod \"398ac086-cb37-46db-a4fa-e964bf468b81\" (UID: \"398ac086-cb37-46db-a4fa-e964bf468b81\") " Sep 30 14:14:41 crc kubenswrapper[4840]: I0930 14:14:41.621385 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/398ac086-cb37-46db-a4fa-e964bf468b81-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "398ac086-cb37-46db-a4fa-e964bf468b81" (UID: "398ac086-cb37-46db-a4fa-e964bf468b81"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:14:41 crc kubenswrapper[4840]: I0930 14:14:41.621421 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/398ac086-cb37-46db-a4fa-e964bf468b81-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "398ac086-cb37-46db-a4fa-e964bf468b81" (UID: "398ac086-cb37-46db-a4fa-e964bf468b81"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:14:41 crc kubenswrapper[4840]: I0930 14:14:41.621705 4840 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/398ac086-cb37-46db-a4fa-e964bf468b81-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:41 crc kubenswrapper[4840]: I0930 14:14:41.621733 4840 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/398ac086-cb37-46db-a4fa-e964bf468b81-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:41 crc kubenswrapper[4840]: I0930 14:14:41.626333 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/398ac086-cb37-46db-a4fa-e964bf468b81-scripts" (OuterVolumeSpecName: "scripts") pod "398ac086-cb37-46db-a4fa-e964bf468b81" (UID: "398ac086-cb37-46db-a4fa-e964bf468b81"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:14:41 crc kubenswrapper[4840]: I0930 14:14:41.629482 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/398ac086-cb37-46db-a4fa-e964bf468b81-kube-api-access-2h7xh" (OuterVolumeSpecName: "kube-api-access-2h7xh") pod "398ac086-cb37-46db-a4fa-e964bf468b81" (UID: "398ac086-cb37-46db-a4fa-e964bf468b81"). InnerVolumeSpecName "kube-api-access-2h7xh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:14:41 crc kubenswrapper[4840]: I0930 14:14:41.649347 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/398ac086-cb37-46db-a4fa-e964bf468b81-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "398ac086-cb37-46db-a4fa-e964bf468b81" (UID: "398ac086-cb37-46db-a4fa-e964bf468b81"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:14:41 crc kubenswrapper[4840]: I0930 14:14:41.715489 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/398ac086-cb37-46db-a4fa-e964bf468b81-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "398ac086-cb37-46db-a4fa-e964bf468b81" (UID: "398ac086-cb37-46db-a4fa-e964bf468b81"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:14:41 crc kubenswrapper[4840]: I0930 14:14:41.723296 4840 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/398ac086-cb37-46db-a4fa-e964bf468b81-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:41 crc kubenswrapper[4840]: I0930 14:14:41.723319 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/398ac086-cb37-46db-a4fa-e964bf468b81-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:41 crc kubenswrapper[4840]: I0930 14:14:41.723330 4840 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/398ac086-cb37-46db-a4fa-e964bf468b81-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:41 crc kubenswrapper[4840]: I0930 14:14:41.723343 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2h7xh\" (UniqueName: \"kubernetes.io/projected/398ac086-cb37-46db-a4fa-e964bf468b81-kube-api-access-2h7xh\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:41 crc kubenswrapper[4840]: I0930 14:14:41.745964 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/398ac086-cb37-46db-a4fa-e964bf468b81-config-data" (OuterVolumeSpecName: "config-data") pod "398ac086-cb37-46db-a4fa-e964bf468b81" (UID: "398ac086-cb37-46db-a4fa-e964bf468b81"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:14:41 crc kubenswrapper[4840]: I0930 14:14:41.824721 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/398ac086-cb37-46db-a4fa-e964bf468b81-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:42 crc kubenswrapper[4840]: I0930 14:14:42.439404 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:14:42 crc kubenswrapper[4840]: I0930 14:14:42.468371 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:14:42 crc kubenswrapper[4840]: I0930 14:14:42.476163 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:14:42 crc kubenswrapper[4840]: I0930 14:14:42.490376 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:14:42 crc kubenswrapper[4840]: E0930 14:14:42.490811 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="398ac086-cb37-46db-a4fa-e964bf468b81" containerName="sg-core" Sep 30 14:14:42 crc kubenswrapper[4840]: I0930 14:14:42.490832 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="398ac086-cb37-46db-a4fa-e964bf468b81" containerName="sg-core" Sep 30 14:14:42 crc kubenswrapper[4840]: E0930 14:14:42.490855 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="398ac086-cb37-46db-a4fa-e964bf468b81" containerName="ceilometer-notification-agent" Sep 30 14:14:42 crc kubenswrapper[4840]: I0930 14:14:42.490863 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="398ac086-cb37-46db-a4fa-e964bf468b81" containerName="ceilometer-notification-agent" Sep 30 14:14:42 crc kubenswrapper[4840]: E0930 14:14:42.490884 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="398ac086-cb37-46db-a4fa-e964bf468b81" containerName="ceilometer-central-agent" Sep 30 14:14:42 crc kubenswrapper[4840]: I0930 14:14:42.490890 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="398ac086-cb37-46db-a4fa-e964bf468b81" containerName="ceilometer-central-agent" Sep 30 14:14:42 crc kubenswrapper[4840]: E0930 14:14:42.490902 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="398ac086-cb37-46db-a4fa-e964bf468b81" containerName="proxy-httpd" Sep 30 14:14:42 crc kubenswrapper[4840]: I0930 14:14:42.490909 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="398ac086-cb37-46db-a4fa-e964bf468b81" containerName="proxy-httpd" Sep 30 14:14:42 crc kubenswrapper[4840]: I0930 14:14:42.491067 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="398ac086-cb37-46db-a4fa-e964bf468b81" containerName="ceilometer-central-agent" Sep 30 14:14:42 crc kubenswrapper[4840]: I0930 14:14:42.491088 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="398ac086-cb37-46db-a4fa-e964bf468b81" containerName="proxy-httpd" Sep 30 14:14:42 crc kubenswrapper[4840]: I0930 14:14:42.491101 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="398ac086-cb37-46db-a4fa-e964bf468b81" containerName="sg-core" Sep 30 14:14:42 crc kubenswrapper[4840]: I0930 14:14:42.491113 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="398ac086-cb37-46db-a4fa-e964bf468b81" containerName="ceilometer-notification-agent" Sep 30 14:14:42 crc kubenswrapper[4840]: I0930 14:14:42.493194 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:14:42 crc kubenswrapper[4840]: I0930 14:14:42.499203 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 30 14:14:42 crc kubenswrapper[4840]: I0930 14:14:42.499220 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 30 14:14:42 crc kubenswrapper[4840]: I0930 14:14:42.506804 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:14:42 crc kubenswrapper[4840]: I0930 14:14:42.637569 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/712feb06-cb97-425d-8b75-fcd5815a94ab-log-httpd\") pod \"ceilometer-0\" (UID: \"712feb06-cb97-425d-8b75-fcd5815a94ab\") " pod="openstack/ceilometer-0" Sep 30 14:14:42 crc kubenswrapper[4840]: I0930 14:14:42.637627 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/712feb06-cb97-425d-8b75-fcd5815a94ab-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"712feb06-cb97-425d-8b75-fcd5815a94ab\") " pod="openstack/ceilometer-0" Sep 30 14:14:42 crc kubenswrapper[4840]: I0930 14:14:42.637662 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/712feb06-cb97-425d-8b75-fcd5815a94ab-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"712feb06-cb97-425d-8b75-fcd5815a94ab\") " pod="openstack/ceilometer-0" Sep 30 14:14:42 crc kubenswrapper[4840]: I0930 14:14:42.637732 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/712feb06-cb97-425d-8b75-fcd5815a94ab-run-httpd\") pod \"ceilometer-0\" (UID: \"712feb06-cb97-425d-8b75-fcd5815a94ab\") " pod="openstack/ceilometer-0" Sep 30 14:14:42 crc kubenswrapper[4840]: I0930 14:14:42.637814 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsh6k\" (UniqueName: \"kubernetes.io/projected/712feb06-cb97-425d-8b75-fcd5815a94ab-kube-api-access-rsh6k\") pod \"ceilometer-0\" (UID: \"712feb06-cb97-425d-8b75-fcd5815a94ab\") " pod="openstack/ceilometer-0" Sep 30 14:14:42 crc kubenswrapper[4840]: I0930 14:14:42.637831 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/712feb06-cb97-425d-8b75-fcd5815a94ab-scripts\") pod \"ceilometer-0\" (UID: \"712feb06-cb97-425d-8b75-fcd5815a94ab\") " pod="openstack/ceilometer-0" Sep 30 14:14:42 crc kubenswrapper[4840]: I0930 14:14:42.637868 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/712feb06-cb97-425d-8b75-fcd5815a94ab-config-data\") pod \"ceilometer-0\" (UID: \"712feb06-cb97-425d-8b75-fcd5815a94ab\") " pod="openstack/ceilometer-0" Sep 30 14:14:42 crc kubenswrapper[4840]: I0930 14:14:42.739702 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/712feb06-cb97-425d-8b75-fcd5815a94ab-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"712feb06-cb97-425d-8b75-fcd5815a94ab\") " pod="openstack/ceilometer-0" Sep 30 14:14:42 crc kubenswrapper[4840]: I0930 14:14:42.739783 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/712feb06-cb97-425d-8b75-fcd5815a94ab-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"712feb06-cb97-425d-8b75-fcd5815a94ab\") " pod="openstack/ceilometer-0" Sep 30 14:14:42 crc kubenswrapper[4840]: I0930 14:14:42.739838 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/712feb06-cb97-425d-8b75-fcd5815a94ab-run-httpd\") pod \"ceilometer-0\" (UID: \"712feb06-cb97-425d-8b75-fcd5815a94ab\") " pod="openstack/ceilometer-0" Sep 30 14:14:42 crc kubenswrapper[4840]: I0930 14:14:42.739914 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsh6k\" (UniqueName: \"kubernetes.io/projected/712feb06-cb97-425d-8b75-fcd5815a94ab-kube-api-access-rsh6k\") pod \"ceilometer-0\" (UID: \"712feb06-cb97-425d-8b75-fcd5815a94ab\") " pod="openstack/ceilometer-0" Sep 30 14:14:42 crc kubenswrapper[4840]: I0930 14:14:42.739932 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/712feb06-cb97-425d-8b75-fcd5815a94ab-scripts\") pod \"ceilometer-0\" (UID: \"712feb06-cb97-425d-8b75-fcd5815a94ab\") " pod="openstack/ceilometer-0" Sep 30 14:14:42 crc kubenswrapper[4840]: I0930 14:14:42.739986 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/712feb06-cb97-425d-8b75-fcd5815a94ab-config-data\") pod \"ceilometer-0\" (UID: \"712feb06-cb97-425d-8b75-fcd5815a94ab\") " pod="openstack/ceilometer-0" Sep 30 14:14:42 crc kubenswrapper[4840]: I0930 14:14:42.740062 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/712feb06-cb97-425d-8b75-fcd5815a94ab-log-httpd\") pod \"ceilometer-0\" (UID: \"712feb06-cb97-425d-8b75-fcd5815a94ab\") " pod="openstack/ceilometer-0" Sep 30 14:14:42 crc kubenswrapper[4840]: I0930 14:14:42.740271 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/712feb06-cb97-425d-8b75-fcd5815a94ab-run-httpd\") pod \"ceilometer-0\" (UID: \"712feb06-cb97-425d-8b75-fcd5815a94ab\") " pod="openstack/ceilometer-0" Sep 30 14:14:42 crc kubenswrapper[4840]: I0930 14:14:42.741132 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/712feb06-cb97-425d-8b75-fcd5815a94ab-log-httpd\") pod \"ceilometer-0\" (UID: \"712feb06-cb97-425d-8b75-fcd5815a94ab\") " pod="openstack/ceilometer-0" Sep 30 14:14:42 crc kubenswrapper[4840]: I0930 14:14:42.744246 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/712feb06-cb97-425d-8b75-fcd5815a94ab-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"712feb06-cb97-425d-8b75-fcd5815a94ab\") " pod="openstack/ceilometer-0" Sep 30 14:14:42 crc kubenswrapper[4840]: I0930 14:14:42.753696 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/712feb06-cb97-425d-8b75-fcd5815a94ab-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"712feb06-cb97-425d-8b75-fcd5815a94ab\") " pod="openstack/ceilometer-0" Sep 30 14:14:42 crc kubenswrapper[4840]: I0930 14:14:42.753828 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/712feb06-cb97-425d-8b75-fcd5815a94ab-config-data\") pod \"ceilometer-0\" (UID: \"712feb06-cb97-425d-8b75-fcd5815a94ab\") " pod="openstack/ceilometer-0" Sep 30 14:14:42 crc kubenswrapper[4840]: I0930 14:14:42.755179 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/712feb06-cb97-425d-8b75-fcd5815a94ab-scripts\") pod \"ceilometer-0\" (UID: \"712feb06-cb97-425d-8b75-fcd5815a94ab\") " pod="openstack/ceilometer-0" Sep 30 14:14:42 crc kubenswrapper[4840]: I0930 14:14:42.756687 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsh6k\" (UniqueName: \"kubernetes.io/projected/712feb06-cb97-425d-8b75-fcd5815a94ab-kube-api-access-rsh6k\") pod \"ceilometer-0\" (UID: \"712feb06-cb97-425d-8b75-fcd5815a94ab\") " pod="openstack/ceilometer-0" Sep 30 14:14:42 crc kubenswrapper[4840]: I0930 14:14:42.822135 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:14:43 crc kubenswrapper[4840]: I0930 14:14:43.254966 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:14:43 crc kubenswrapper[4840]: I0930 14:14:43.449392 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"712feb06-cb97-425d-8b75-fcd5815a94ab","Type":"ContainerStarted","Data":"a0ed70585198d5191ed48940c56b05d8a544bb56844542b03b3b43d683eb138f"} Sep 30 14:14:44 crc kubenswrapper[4840]: I0930 14:14:44.128790 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="398ac086-cb37-46db-a4fa-e964bf468b81" path="/var/lib/kubelet/pods/398ac086-cb37-46db-a4fa-e964bf468b81/volumes" Sep 30 14:14:46 crc kubenswrapper[4840]: I0930 14:14:46.476519 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"712feb06-cb97-425d-8b75-fcd5815a94ab","Type":"ContainerStarted","Data":"9ffdf9131bd8af23b5cc76199769f964619ded52d0516748629ec0e503506d3c"} Sep 30 14:14:47 crc kubenswrapper[4840]: I0930 14:14:47.485685 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"712feb06-cb97-425d-8b75-fcd5815a94ab","Type":"ContainerStarted","Data":"fa186597d4025b0b4a5b61ec12ce99d1733596a1825f242ee2bcbfc59b2000c6"} Sep 30 14:14:48 crc kubenswrapper[4840]: I0930 14:14:48.495015 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"712feb06-cb97-425d-8b75-fcd5815a94ab","Type":"ContainerStarted","Data":"53aa98fbc16bc77ba4f16d0623177aa092f9f3f0a29df3651490494fd3510c5d"} Sep 30 14:14:50 crc kubenswrapper[4840]: I0930 14:14:50.514454 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"712feb06-cb97-425d-8b75-fcd5815a94ab","Type":"ContainerStarted","Data":"25e57148f578d9ab12ef79020c7ac3c16f6b8542b50b2833a6d175c2d8a4c3f0"} Sep 30 14:14:50 crc kubenswrapper[4840]: I0930 14:14:50.515105 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 30 14:14:50 crc kubenswrapper[4840]: I0930 14:14:50.539698 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.497287318 podStartE2EDuration="8.539673767s" podCreationTimestamp="2025-09-30 14:14:42 +0000 UTC" firstStartedPulling="2025-09-30 14:14:43.262603608 +0000 UTC m=+1111.891690031" lastFinishedPulling="2025-09-30 14:14:49.304990057 +0000 UTC m=+1117.934076480" observedRunningTime="2025-09-30 14:14:50.532658544 +0000 UTC m=+1119.161744977" watchObservedRunningTime="2025-09-30 14:14:50.539673767 +0000 UTC m=+1119.168760190" Sep 30 14:14:54 crc kubenswrapper[4840]: I0930 14:14:54.547694 4840 generic.go:334] "Generic (PLEG): container finished" podID="d7e57e20-fd49-4406-9814-1486cc9efdf0" containerID="ea758c4f9027095d42744a11ef0b508192d6c55fb24d66b7de02b3df32145dcf" exitCode=0 Sep 30 14:14:54 crc kubenswrapper[4840]: I0930 14:14:54.547775 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-nlp8v" event={"ID":"d7e57e20-fd49-4406-9814-1486cc9efdf0","Type":"ContainerDied","Data":"ea758c4f9027095d42744a11ef0b508192d6c55fb24d66b7de02b3df32145dcf"} Sep 30 14:14:55 crc kubenswrapper[4840]: I0930 14:14:55.914576 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-nlp8v" Sep 30 14:14:55 crc kubenswrapper[4840]: I0930 14:14:55.979443 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhqw8\" (UniqueName: \"kubernetes.io/projected/d7e57e20-fd49-4406-9814-1486cc9efdf0-kube-api-access-jhqw8\") pod \"d7e57e20-fd49-4406-9814-1486cc9efdf0\" (UID: \"d7e57e20-fd49-4406-9814-1486cc9efdf0\") " Sep 30 14:14:55 crc kubenswrapper[4840]: I0930 14:14:55.979615 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7e57e20-fd49-4406-9814-1486cc9efdf0-combined-ca-bundle\") pod \"d7e57e20-fd49-4406-9814-1486cc9efdf0\" (UID: \"d7e57e20-fd49-4406-9814-1486cc9efdf0\") " Sep 30 14:14:55 crc kubenswrapper[4840]: I0930 14:14:55.979747 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7e57e20-fd49-4406-9814-1486cc9efdf0-scripts\") pod \"d7e57e20-fd49-4406-9814-1486cc9efdf0\" (UID: \"d7e57e20-fd49-4406-9814-1486cc9efdf0\") " Sep 30 14:14:55 crc kubenswrapper[4840]: I0930 14:14:55.979785 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7e57e20-fd49-4406-9814-1486cc9efdf0-config-data\") pod \"d7e57e20-fd49-4406-9814-1486cc9efdf0\" (UID: \"d7e57e20-fd49-4406-9814-1486cc9efdf0\") " Sep 30 14:14:55 crc kubenswrapper[4840]: I0930 14:14:55.985201 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7e57e20-fd49-4406-9814-1486cc9efdf0-scripts" (OuterVolumeSpecName: "scripts") pod "d7e57e20-fd49-4406-9814-1486cc9efdf0" (UID: "d7e57e20-fd49-4406-9814-1486cc9efdf0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:14:55 crc kubenswrapper[4840]: I0930 14:14:55.985367 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7e57e20-fd49-4406-9814-1486cc9efdf0-kube-api-access-jhqw8" (OuterVolumeSpecName: "kube-api-access-jhqw8") pod "d7e57e20-fd49-4406-9814-1486cc9efdf0" (UID: "d7e57e20-fd49-4406-9814-1486cc9efdf0"). InnerVolumeSpecName "kube-api-access-jhqw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:14:56 crc kubenswrapper[4840]: I0930 14:14:56.008043 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7e57e20-fd49-4406-9814-1486cc9efdf0-config-data" (OuterVolumeSpecName: "config-data") pod "d7e57e20-fd49-4406-9814-1486cc9efdf0" (UID: "d7e57e20-fd49-4406-9814-1486cc9efdf0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:14:56 crc kubenswrapper[4840]: I0930 14:14:56.009652 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7e57e20-fd49-4406-9814-1486cc9efdf0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d7e57e20-fd49-4406-9814-1486cc9efdf0" (UID: "d7e57e20-fd49-4406-9814-1486cc9efdf0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:14:56 crc kubenswrapper[4840]: I0930 14:14:56.081455 4840 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7e57e20-fd49-4406-9814-1486cc9efdf0-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:56 crc kubenswrapper[4840]: I0930 14:14:56.081486 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7e57e20-fd49-4406-9814-1486cc9efdf0-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:56 crc kubenswrapper[4840]: I0930 14:14:56.081499 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhqw8\" (UniqueName: \"kubernetes.io/projected/d7e57e20-fd49-4406-9814-1486cc9efdf0-kube-api-access-jhqw8\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:56 crc kubenswrapper[4840]: I0930 14:14:56.081508 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7e57e20-fd49-4406-9814-1486cc9efdf0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:14:56 crc kubenswrapper[4840]: I0930 14:14:56.565799 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-nlp8v" event={"ID":"d7e57e20-fd49-4406-9814-1486cc9efdf0","Type":"ContainerDied","Data":"77237d21476f98c7971b9237c78482ec28734e1382088face1e615691cd26659"} Sep 30 14:14:56 crc kubenswrapper[4840]: I0930 14:14:56.565843 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="77237d21476f98c7971b9237c78482ec28734e1382088face1e615691cd26659" Sep 30 14:14:56 crc kubenswrapper[4840]: I0930 14:14:56.565853 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-nlp8v" Sep 30 14:14:56 crc kubenswrapper[4840]: I0930 14:14:56.661952 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Sep 30 14:14:56 crc kubenswrapper[4840]: E0930 14:14:56.662685 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7e57e20-fd49-4406-9814-1486cc9efdf0" containerName="nova-cell0-conductor-db-sync" Sep 30 14:14:56 crc kubenswrapper[4840]: I0930 14:14:56.662705 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7e57e20-fd49-4406-9814-1486cc9efdf0" containerName="nova-cell0-conductor-db-sync" Sep 30 14:14:56 crc kubenswrapper[4840]: I0930 14:14:56.662878 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7e57e20-fd49-4406-9814-1486cc9efdf0" containerName="nova-cell0-conductor-db-sync" Sep 30 14:14:56 crc kubenswrapper[4840]: I0930 14:14:56.663448 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Sep 30 14:14:56 crc kubenswrapper[4840]: I0930 14:14:56.670051 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Sep 30 14:14:56 crc kubenswrapper[4840]: I0930 14:14:56.670253 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-bgvhh" Sep 30 14:14:56 crc kubenswrapper[4840]: I0930 14:14:56.679818 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Sep 30 14:14:56 crc kubenswrapper[4840]: I0930 14:14:56.799753 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/739533f4-30e5-4941-a7f3-d43a945cd8e8-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"739533f4-30e5-4941-a7f3-d43a945cd8e8\") " pod="openstack/nova-cell0-conductor-0" Sep 30 14:14:56 crc kubenswrapper[4840]: I0930 14:14:56.799809 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6pn5\" (UniqueName: \"kubernetes.io/projected/739533f4-30e5-4941-a7f3-d43a945cd8e8-kube-api-access-z6pn5\") pod \"nova-cell0-conductor-0\" (UID: \"739533f4-30e5-4941-a7f3-d43a945cd8e8\") " pod="openstack/nova-cell0-conductor-0" Sep 30 14:14:56 crc kubenswrapper[4840]: I0930 14:14:56.800071 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/739533f4-30e5-4941-a7f3-d43a945cd8e8-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"739533f4-30e5-4941-a7f3-d43a945cd8e8\") " pod="openstack/nova-cell0-conductor-0" Sep 30 14:14:56 crc kubenswrapper[4840]: I0930 14:14:56.901407 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/739533f4-30e5-4941-a7f3-d43a945cd8e8-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"739533f4-30e5-4941-a7f3-d43a945cd8e8\") " pod="openstack/nova-cell0-conductor-0" Sep 30 14:14:56 crc kubenswrapper[4840]: I0930 14:14:56.901539 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/739533f4-30e5-4941-a7f3-d43a945cd8e8-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"739533f4-30e5-4941-a7f3-d43a945cd8e8\") " pod="openstack/nova-cell0-conductor-0" Sep 30 14:14:56 crc kubenswrapper[4840]: I0930 14:14:56.901586 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6pn5\" (UniqueName: \"kubernetes.io/projected/739533f4-30e5-4941-a7f3-d43a945cd8e8-kube-api-access-z6pn5\") pod \"nova-cell0-conductor-0\" (UID: \"739533f4-30e5-4941-a7f3-d43a945cd8e8\") " pod="openstack/nova-cell0-conductor-0" Sep 30 14:14:56 crc kubenswrapper[4840]: I0930 14:14:56.907857 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/739533f4-30e5-4941-a7f3-d43a945cd8e8-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"739533f4-30e5-4941-a7f3-d43a945cd8e8\") " pod="openstack/nova-cell0-conductor-0" Sep 30 14:14:56 crc kubenswrapper[4840]: I0930 14:14:56.909140 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/739533f4-30e5-4941-a7f3-d43a945cd8e8-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"739533f4-30e5-4941-a7f3-d43a945cd8e8\") " pod="openstack/nova-cell0-conductor-0" Sep 30 14:14:56 crc kubenswrapper[4840]: I0930 14:14:56.923260 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6pn5\" (UniqueName: \"kubernetes.io/projected/739533f4-30e5-4941-a7f3-d43a945cd8e8-kube-api-access-z6pn5\") pod \"nova-cell0-conductor-0\" (UID: \"739533f4-30e5-4941-a7f3-d43a945cd8e8\") " pod="openstack/nova-cell0-conductor-0" Sep 30 14:14:56 crc kubenswrapper[4840]: I0930 14:14:56.983658 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Sep 30 14:14:57 crc kubenswrapper[4840]: I0930 14:14:57.422734 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Sep 30 14:14:57 crc kubenswrapper[4840]: I0930 14:14:57.575419 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"739533f4-30e5-4941-a7f3-d43a945cd8e8","Type":"ContainerStarted","Data":"6f5f060c610bacc719f0f5e5bd0528dd25355c377dbee4dd3b5350a9247fb927"} Sep 30 14:14:58 crc kubenswrapper[4840]: I0930 14:14:58.585526 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"739533f4-30e5-4941-a7f3-d43a945cd8e8","Type":"ContainerStarted","Data":"cac6a11246f4b0f5822ef1df3684bd51a4c8d53b4049283df9f8a6a8a77410ff"} Sep 30 14:14:58 crc kubenswrapper[4840]: I0930 14:14:58.585869 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Sep 30 14:14:58 crc kubenswrapper[4840]: I0930 14:14:58.605431 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.60540799 podStartE2EDuration="2.60540799s" podCreationTimestamp="2025-09-30 14:14:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:14:58.600931313 +0000 UTC m=+1127.230017736" watchObservedRunningTime="2025-09-30 14:14:58.60540799 +0000 UTC m=+1127.234494433" Sep 30 14:15:00 crc kubenswrapper[4840]: I0930 14:15:00.144616 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320695-ml6k2"] Sep 30 14:15:00 crc kubenswrapper[4840]: I0930 14:15:00.146448 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320695-ml6k2" Sep 30 14:15:00 crc kubenswrapper[4840]: I0930 14:15:00.148981 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 30 14:15:00 crc kubenswrapper[4840]: I0930 14:15:00.149686 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 30 14:15:00 crc kubenswrapper[4840]: I0930 14:15:00.161320 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320695-ml6k2"] Sep 30 14:15:00 crc kubenswrapper[4840]: I0930 14:15:00.263344 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/af8e3cc6-9628-4d6b-81c6-41a55cfb8935-secret-volume\") pod \"collect-profiles-29320695-ml6k2\" (UID: \"af8e3cc6-9628-4d6b-81c6-41a55cfb8935\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320695-ml6k2" Sep 30 14:15:00 crc kubenswrapper[4840]: I0930 14:15:00.263477 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/af8e3cc6-9628-4d6b-81c6-41a55cfb8935-config-volume\") pod \"collect-profiles-29320695-ml6k2\" (UID: \"af8e3cc6-9628-4d6b-81c6-41a55cfb8935\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320695-ml6k2" Sep 30 14:15:00 crc kubenswrapper[4840]: I0930 14:15:00.263523 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p42qr\" (UniqueName: \"kubernetes.io/projected/af8e3cc6-9628-4d6b-81c6-41a55cfb8935-kube-api-access-p42qr\") pod \"collect-profiles-29320695-ml6k2\" (UID: \"af8e3cc6-9628-4d6b-81c6-41a55cfb8935\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320695-ml6k2" Sep 30 14:15:00 crc kubenswrapper[4840]: I0930 14:15:00.365143 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/af8e3cc6-9628-4d6b-81c6-41a55cfb8935-config-volume\") pod \"collect-profiles-29320695-ml6k2\" (UID: \"af8e3cc6-9628-4d6b-81c6-41a55cfb8935\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320695-ml6k2" Sep 30 14:15:00 crc kubenswrapper[4840]: I0930 14:15:00.365242 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p42qr\" (UniqueName: \"kubernetes.io/projected/af8e3cc6-9628-4d6b-81c6-41a55cfb8935-kube-api-access-p42qr\") pod \"collect-profiles-29320695-ml6k2\" (UID: \"af8e3cc6-9628-4d6b-81c6-41a55cfb8935\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320695-ml6k2" Sep 30 14:15:00 crc kubenswrapper[4840]: I0930 14:15:00.365311 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/af8e3cc6-9628-4d6b-81c6-41a55cfb8935-secret-volume\") pod \"collect-profiles-29320695-ml6k2\" (UID: \"af8e3cc6-9628-4d6b-81c6-41a55cfb8935\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320695-ml6k2" Sep 30 14:15:00 crc kubenswrapper[4840]: I0930 14:15:00.366605 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/af8e3cc6-9628-4d6b-81c6-41a55cfb8935-config-volume\") pod \"collect-profiles-29320695-ml6k2\" (UID: \"af8e3cc6-9628-4d6b-81c6-41a55cfb8935\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320695-ml6k2" Sep 30 14:15:00 crc kubenswrapper[4840]: I0930 14:15:00.371902 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/af8e3cc6-9628-4d6b-81c6-41a55cfb8935-secret-volume\") pod \"collect-profiles-29320695-ml6k2\" (UID: \"af8e3cc6-9628-4d6b-81c6-41a55cfb8935\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320695-ml6k2" Sep 30 14:15:00 crc kubenswrapper[4840]: I0930 14:15:00.390209 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p42qr\" (UniqueName: \"kubernetes.io/projected/af8e3cc6-9628-4d6b-81c6-41a55cfb8935-kube-api-access-p42qr\") pod \"collect-profiles-29320695-ml6k2\" (UID: \"af8e3cc6-9628-4d6b-81c6-41a55cfb8935\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320695-ml6k2" Sep 30 14:15:00 crc kubenswrapper[4840]: I0930 14:15:00.472836 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320695-ml6k2" Sep 30 14:15:00 crc kubenswrapper[4840]: I0930 14:15:00.906685 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320695-ml6k2"] Sep 30 14:15:01 crc kubenswrapper[4840]: I0930 14:15:01.616401 4840 generic.go:334] "Generic (PLEG): container finished" podID="af8e3cc6-9628-4d6b-81c6-41a55cfb8935" containerID="477513fbf915108bacd028dfd28bd853d6596414cf70f5430f7438625ae9e2e4" exitCode=0 Sep 30 14:15:01 crc kubenswrapper[4840]: I0930 14:15:01.616510 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320695-ml6k2" event={"ID":"af8e3cc6-9628-4d6b-81c6-41a55cfb8935","Type":"ContainerDied","Data":"477513fbf915108bacd028dfd28bd853d6596414cf70f5430f7438625ae9e2e4"} Sep 30 14:15:01 crc kubenswrapper[4840]: I0930 14:15:01.616765 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320695-ml6k2" event={"ID":"af8e3cc6-9628-4d6b-81c6-41a55cfb8935","Type":"ContainerStarted","Data":"8a0ab7444e0bbb6c5d36e2afd118738335089afe095939ac4b5fd74819c00221"} Sep 30 14:15:02 crc kubenswrapper[4840]: I0930 14:15:02.954957 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320695-ml6k2" Sep 30 14:15:03 crc kubenswrapper[4840]: I0930 14:15:03.012622 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p42qr\" (UniqueName: \"kubernetes.io/projected/af8e3cc6-9628-4d6b-81c6-41a55cfb8935-kube-api-access-p42qr\") pod \"af8e3cc6-9628-4d6b-81c6-41a55cfb8935\" (UID: \"af8e3cc6-9628-4d6b-81c6-41a55cfb8935\") " Sep 30 14:15:03 crc kubenswrapper[4840]: I0930 14:15:03.012698 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/af8e3cc6-9628-4d6b-81c6-41a55cfb8935-config-volume\") pod \"af8e3cc6-9628-4d6b-81c6-41a55cfb8935\" (UID: \"af8e3cc6-9628-4d6b-81c6-41a55cfb8935\") " Sep 30 14:15:03 crc kubenswrapper[4840]: I0930 14:15:03.012720 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/af8e3cc6-9628-4d6b-81c6-41a55cfb8935-secret-volume\") pod \"af8e3cc6-9628-4d6b-81c6-41a55cfb8935\" (UID: \"af8e3cc6-9628-4d6b-81c6-41a55cfb8935\") " Sep 30 14:15:03 crc kubenswrapper[4840]: I0930 14:15:03.014088 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af8e3cc6-9628-4d6b-81c6-41a55cfb8935-config-volume" (OuterVolumeSpecName: "config-volume") pod "af8e3cc6-9628-4d6b-81c6-41a55cfb8935" (UID: "af8e3cc6-9628-4d6b-81c6-41a55cfb8935"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:15:03 crc kubenswrapper[4840]: I0930 14:15:03.018229 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af8e3cc6-9628-4d6b-81c6-41a55cfb8935-kube-api-access-p42qr" (OuterVolumeSpecName: "kube-api-access-p42qr") pod "af8e3cc6-9628-4d6b-81c6-41a55cfb8935" (UID: "af8e3cc6-9628-4d6b-81c6-41a55cfb8935"). InnerVolumeSpecName "kube-api-access-p42qr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:15:03 crc kubenswrapper[4840]: I0930 14:15:03.018796 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af8e3cc6-9628-4d6b-81c6-41a55cfb8935-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "af8e3cc6-9628-4d6b-81c6-41a55cfb8935" (UID: "af8e3cc6-9628-4d6b-81c6-41a55cfb8935"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:15:03 crc kubenswrapper[4840]: I0930 14:15:03.114965 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p42qr\" (UniqueName: \"kubernetes.io/projected/af8e3cc6-9628-4d6b-81c6-41a55cfb8935-kube-api-access-p42qr\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:03 crc kubenswrapper[4840]: I0930 14:15:03.114996 4840 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/af8e3cc6-9628-4d6b-81c6-41a55cfb8935-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:03 crc kubenswrapper[4840]: I0930 14:15:03.115005 4840 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/af8e3cc6-9628-4d6b-81c6-41a55cfb8935-config-volume\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:03 crc kubenswrapper[4840]: I0930 14:15:03.637174 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320695-ml6k2" event={"ID":"af8e3cc6-9628-4d6b-81c6-41a55cfb8935","Type":"ContainerDied","Data":"8a0ab7444e0bbb6c5d36e2afd118738335089afe095939ac4b5fd74819c00221"} Sep 30 14:15:03 crc kubenswrapper[4840]: I0930 14:15:03.637387 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8a0ab7444e0bbb6c5d36e2afd118738335089afe095939ac4b5fd74819c00221" Sep 30 14:15:03 crc kubenswrapper[4840]: I0930 14:15:03.637244 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320695-ml6k2" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.012718 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.432941 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-v9q8t"] Sep 30 14:15:07 crc kubenswrapper[4840]: E0930 14:15:07.433480 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af8e3cc6-9628-4d6b-81c6-41a55cfb8935" containerName="collect-profiles" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.433507 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="af8e3cc6-9628-4d6b-81c6-41a55cfb8935" containerName="collect-profiles" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.433761 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="af8e3cc6-9628-4d6b-81c6-41a55cfb8935" containerName="collect-profiles" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.434566 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-v9q8t" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.439074 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.439495 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.453877 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-v9q8t"] Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.501357 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/722796b2-e357-4fec-8437-46f163f510d6-scripts\") pod \"nova-cell0-cell-mapping-v9q8t\" (UID: \"722796b2-e357-4fec-8437-46f163f510d6\") " pod="openstack/nova-cell0-cell-mapping-v9q8t" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.501445 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/722796b2-e357-4fec-8437-46f163f510d6-config-data\") pod \"nova-cell0-cell-mapping-v9q8t\" (UID: \"722796b2-e357-4fec-8437-46f163f510d6\") " pod="openstack/nova-cell0-cell-mapping-v9q8t" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.501508 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxrxr\" (UniqueName: \"kubernetes.io/projected/722796b2-e357-4fec-8437-46f163f510d6-kube-api-access-mxrxr\") pod \"nova-cell0-cell-mapping-v9q8t\" (UID: \"722796b2-e357-4fec-8437-46f163f510d6\") " pod="openstack/nova-cell0-cell-mapping-v9q8t" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.501638 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/722796b2-e357-4fec-8437-46f163f510d6-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-v9q8t\" (UID: \"722796b2-e357-4fec-8437-46f163f510d6\") " pod="openstack/nova-cell0-cell-mapping-v9q8t" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.583118 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.585028 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.595415 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.596118 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.602941 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/722796b2-e357-4fec-8437-46f163f510d6-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-v9q8t\" (UID: \"722796b2-e357-4fec-8437-46f163f510d6\") " pod="openstack/nova-cell0-cell-mapping-v9q8t" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.603004 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/722796b2-e357-4fec-8437-46f163f510d6-scripts\") pod \"nova-cell0-cell-mapping-v9q8t\" (UID: \"722796b2-e357-4fec-8437-46f163f510d6\") " pod="openstack/nova-cell0-cell-mapping-v9q8t" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.603060 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/722796b2-e357-4fec-8437-46f163f510d6-config-data\") pod \"nova-cell0-cell-mapping-v9q8t\" (UID: \"722796b2-e357-4fec-8437-46f163f510d6\") " pod="openstack/nova-cell0-cell-mapping-v9q8t" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.603099 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxrxr\" (UniqueName: \"kubernetes.io/projected/722796b2-e357-4fec-8437-46f163f510d6-kube-api-access-mxrxr\") pod \"nova-cell0-cell-mapping-v9q8t\" (UID: \"722796b2-e357-4fec-8437-46f163f510d6\") " pod="openstack/nova-cell0-cell-mapping-v9q8t" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.613631 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/722796b2-e357-4fec-8437-46f163f510d6-scripts\") pod \"nova-cell0-cell-mapping-v9q8t\" (UID: \"722796b2-e357-4fec-8437-46f163f510d6\") " pod="openstack/nova-cell0-cell-mapping-v9q8t" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.614209 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/722796b2-e357-4fec-8437-46f163f510d6-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-v9q8t\" (UID: \"722796b2-e357-4fec-8437-46f163f510d6\") " pod="openstack/nova-cell0-cell-mapping-v9q8t" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.630381 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/722796b2-e357-4fec-8437-46f163f510d6-config-data\") pod \"nova-cell0-cell-mapping-v9q8t\" (UID: \"722796b2-e357-4fec-8437-46f163f510d6\") " pod="openstack/nova-cell0-cell-mapping-v9q8t" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.634356 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxrxr\" (UniqueName: \"kubernetes.io/projected/722796b2-e357-4fec-8437-46f163f510d6-kube-api-access-mxrxr\") pod \"nova-cell0-cell-mapping-v9q8t\" (UID: \"722796b2-e357-4fec-8437-46f163f510d6\") " pod="openstack/nova-cell0-cell-mapping-v9q8t" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.664620 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.670577 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.715248 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.753326 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc04dbcf-177f-4039-be37-9986232e2d42-config-data\") pod \"nova-api-0\" (UID: \"bc04dbcf-177f-4039-be37-9986232e2d42\") " pod="openstack/nova-api-0" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.753642 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc04dbcf-177f-4039-be37-9986232e2d42-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"bc04dbcf-177f-4039-be37-9986232e2d42\") " pod="openstack/nova-api-0" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.754062 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2kz5\" (UniqueName: \"kubernetes.io/projected/bc04dbcf-177f-4039-be37-9986232e2d42-kube-api-access-h2kz5\") pod \"nova-api-0\" (UID: \"bc04dbcf-177f-4039-be37-9986232e2d42\") " pod="openstack/nova-api-0" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.754467 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc04dbcf-177f-4039-be37-9986232e2d42-logs\") pod \"nova-api-0\" (UID: \"bc04dbcf-177f-4039-be37-9986232e2d42\") " pod="openstack/nova-api-0" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.761342 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.791344 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-v9q8t" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.856127 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70d820fe-33fc-4d18-b75f-3bff2044ba5b-logs\") pod \"nova-metadata-0\" (UID: \"70d820fe-33fc-4d18-b75f-3bff2044ba5b\") " pod="openstack/nova-metadata-0" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.856202 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2kz5\" (UniqueName: \"kubernetes.io/projected/bc04dbcf-177f-4039-be37-9986232e2d42-kube-api-access-h2kz5\") pod \"nova-api-0\" (UID: \"bc04dbcf-177f-4039-be37-9986232e2d42\") " pod="openstack/nova-api-0" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.856226 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc04dbcf-177f-4039-be37-9986232e2d42-logs\") pod \"nova-api-0\" (UID: \"bc04dbcf-177f-4039-be37-9986232e2d42\") " pod="openstack/nova-api-0" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.856249 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fw4kk\" (UniqueName: \"kubernetes.io/projected/70d820fe-33fc-4d18-b75f-3bff2044ba5b-kube-api-access-fw4kk\") pod \"nova-metadata-0\" (UID: \"70d820fe-33fc-4d18-b75f-3bff2044ba5b\") " pod="openstack/nova-metadata-0" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.856338 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70d820fe-33fc-4d18-b75f-3bff2044ba5b-config-data\") pod \"nova-metadata-0\" (UID: \"70d820fe-33fc-4d18-b75f-3bff2044ba5b\") " pod="openstack/nova-metadata-0" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.856390 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc04dbcf-177f-4039-be37-9986232e2d42-config-data\") pod \"nova-api-0\" (UID: \"bc04dbcf-177f-4039-be37-9986232e2d42\") " pod="openstack/nova-api-0" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.856412 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc04dbcf-177f-4039-be37-9986232e2d42-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"bc04dbcf-177f-4039-be37-9986232e2d42\") " pod="openstack/nova-api-0" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.856450 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70d820fe-33fc-4d18-b75f-3bff2044ba5b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"70d820fe-33fc-4d18-b75f-3bff2044ba5b\") " pod="openstack/nova-metadata-0" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.857311 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc04dbcf-177f-4039-be37-9986232e2d42-logs\") pod \"nova-api-0\" (UID: \"bc04dbcf-177f-4039-be37-9986232e2d42\") " pod="openstack/nova-api-0" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.880544 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc04dbcf-177f-4039-be37-9986232e2d42-config-data\") pod \"nova-api-0\" (UID: \"bc04dbcf-177f-4039-be37-9986232e2d42\") " pod="openstack/nova-api-0" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.896483 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc04dbcf-177f-4039-be37-9986232e2d42-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"bc04dbcf-177f-4039-be37-9986232e2d42\") " pod="openstack/nova-api-0" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.896574 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.898030 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.902068 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.934267 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2kz5\" (UniqueName: \"kubernetes.io/projected/bc04dbcf-177f-4039-be37-9986232e2d42-kube-api-access-h2kz5\") pod \"nova-api-0\" (UID: \"bc04dbcf-177f-4039-be37-9986232e2d42\") " pod="openstack/nova-api-0" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.952619 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.962084 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e616a00a-2b1e-4ea6-bde6-1a5517c8464e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e616a00a-2b1e-4ea6-bde6-1a5517c8464e\") " pod="openstack/nova-scheduler-0" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.962180 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70d820fe-33fc-4d18-b75f-3bff2044ba5b-logs\") pod \"nova-metadata-0\" (UID: \"70d820fe-33fc-4d18-b75f-3bff2044ba5b\") " pod="openstack/nova-metadata-0" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.962226 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fw4kk\" (UniqueName: \"kubernetes.io/projected/70d820fe-33fc-4d18-b75f-3bff2044ba5b-kube-api-access-fw4kk\") pod \"nova-metadata-0\" (UID: \"70d820fe-33fc-4d18-b75f-3bff2044ba5b\") " pod="openstack/nova-metadata-0" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.962267 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjxwk\" (UniqueName: \"kubernetes.io/projected/e616a00a-2b1e-4ea6-bde6-1a5517c8464e-kube-api-access-sjxwk\") pod \"nova-scheduler-0\" (UID: \"e616a00a-2b1e-4ea6-bde6-1a5517c8464e\") " pod="openstack/nova-scheduler-0" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.962304 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e616a00a-2b1e-4ea6-bde6-1a5517c8464e-config-data\") pod \"nova-scheduler-0\" (UID: \"e616a00a-2b1e-4ea6-bde6-1a5517c8464e\") " pod="openstack/nova-scheduler-0" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.962336 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70d820fe-33fc-4d18-b75f-3bff2044ba5b-config-data\") pod \"nova-metadata-0\" (UID: \"70d820fe-33fc-4d18-b75f-3bff2044ba5b\") " pod="openstack/nova-metadata-0" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.962400 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70d820fe-33fc-4d18-b75f-3bff2044ba5b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"70d820fe-33fc-4d18-b75f-3bff2044ba5b\") " pod="openstack/nova-metadata-0" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.969274 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70d820fe-33fc-4d18-b75f-3bff2044ba5b-logs\") pod \"nova-metadata-0\" (UID: \"70d820fe-33fc-4d18-b75f-3bff2044ba5b\") " pod="openstack/nova-metadata-0" Sep 30 14:15:07 crc kubenswrapper[4840]: I0930 14:15:07.974189 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70d820fe-33fc-4d18-b75f-3bff2044ba5b-config-data\") pod \"nova-metadata-0\" (UID: \"70d820fe-33fc-4d18-b75f-3bff2044ba5b\") " pod="openstack/nova-metadata-0" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.008409 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-kgxlg"] Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.010354 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-kgxlg" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.014442 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fw4kk\" (UniqueName: \"kubernetes.io/projected/70d820fe-33fc-4d18-b75f-3bff2044ba5b-kube-api-access-fw4kk\") pod \"nova-metadata-0\" (UID: \"70d820fe-33fc-4d18-b75f-3bff2044ba5b\") " pod="openstack/nova-metadata-0" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.035300 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70d820fe-33fc-4d18-b75f-3bff2044ba5b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"70d820fe-33fc-4d18-b75f-3bff2044ba5b\") " pod="openstack/nova-metadata-0" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.051624 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.053325 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.055871 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.062262 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-kgxlg"] Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.066927 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/85b94340-ea1f-4cde-b359-d4970084e6f5-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-kgxlg\" (UID: \"85b94340-ea1f-4cde-b359-d4970084e6f5\") " pod="openstack/dnsmasq-dns-845d6d6f59-kgxlg" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.067002 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/85b94340-ea1f-4cde-b359-d4970084e6f5-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-kgxlg\" (UID: \"85b94340-ea1f-4cde-b359-d4970084e6f5\") " pod="openstack/dnsmasq-dns-845d6d6f59-kgxlg" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.067040 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e616a00a-2b1e-4ea6-bde6-1a5517c8464e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e616a00a-2b1e-4ea6-bde6-1a5517c8464e\") " pod="openstack/nova-scheduler-0" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.067116 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/85b94340-ea1f-4cde-b359-d4970084e6f5-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-kgxlg\" (UID: \"85b94340-ea1f-4cde-b359-d4970084e6f5\") " pod="openstack/dnsmasq-dns-845d6d6f59-kgxlg" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.067151 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqxjp\" (UniqueName: \"kubernetes.io/projected/85b94340-ea1f-4cde-b359-d4970084e6f5-kube-api-access-rqxjp\") pod \"dnsmasq-dns-845d6d6f59-kgxlg\" (UID: \"85b94340-ea1f-4cde-b359-d4970084e6f5\") " pod="openstack/dnsmasq-dns-845d6d6f59-kgxlg" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.067181 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85b94340-ea1f-4cde-b359-d4970084e6f5-config\") pod \"dnsmasq-dns-845d6d6f59-kgxlg\" (UID: \"85b94340-ea1f-4cde-b359-d4970084e6f5\") " pod="openstack/dnsmasq-dns-845d6d6f59-kgxlg" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.067221 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjxwk\" (UniqueName: \"kubernetes.io/projected/e616a00a-2b1e-4ea6-bde6-1a5517c8464e-kube-api-access-sjxwk\") pod \"nova-scheduler-0\" (UID: \"e616a00a-2b1e-4ea6-bde6-1a5517c8464e\") " pod="openstack/nova-scheduler-0" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.067249 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e616a00a-2b1e-4ea6-bde6-1a5517c8464e-config-data\") pod \"nova-scheduler-0\" (UID: \"e616a00a-2b1e-4ea6-bde6-1a5517c8464e\") " pod="openstack/nova-scheduler-0" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.067269 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/85b94340-ea1f-4cde-b359-d4970084e6f5-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-kgxlg\" (UID: \"85b94340-ea1f-4cde-b359-d4970084e6f5\") " pod="openstack/dnsmasq-dns-845d6d6f59-kgxlg" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.072881 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.075603 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e616a00a-2b1e-4ea6-bde6-1a5517c8464e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e616a00a-2b1e-4ea6-bde6-1a5517c8464e\") " pod="openstack/nova-scheduler-0" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.082184 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e616a00a-2b1e-4ea6-bde6-1a5517c8464e-config-data\") pod \"nova-scheduler-0\" (UID: \"e616a00a-2b1e-4ea6-bde6-1a5517c8464e\") " pod="openstack/nova-scheduler-0" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.102890 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjxwk\" (UniqueName: \"kubernetes.io/projected/e616a00a-2b1e-4ea6-bde6-1a5517c8464e-kube-api-access-sjxwk\") pod \"nova-scheduler-0\" (UID: \"e616a00a-2b1e-4ea6-bde6-1a5517c8464e\") " pod="openstack/nova-scheduler-0" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.170257 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.193197 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/85b94340-ea1f-4cde-b359-d4970084e6f5-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-kgxlg\" (UID: \"85b94340-ea1f-4cde-b359-d4970084e6f5\") " pod="openstack/dnsmasq-dns-845d6d6f59-kgxlg" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.193251 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqxjp\" (UniqueName: \"kubernetes.io/projected/85b94340-ea1f-4cde-b359-d4970084e6f5-kube-api-access-rqxjp\") pod \"dnsmasq-dns-845d6d6f59-kgxlg\" (UID: \"85b94340-ea1f-4cde-b359-d4970084e6f5\") " pod="openstack/dnsmasq-dns-845d6d6f59-kgxlg" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.193291 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85b94340-ea1f-4cde-b359-d4970084e6f5-config\") pod \"dnsmasq-dns-845d6d6f59-kgxlg\" (UID: \"85b94340-ea1f-4cde-b359-d4970084e6f5\") " pod="openstack/dnsmasq-dns-845d6d6f59-kgxlg" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.193401 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/85b94340-ea1f-4cde-b359-d4970084e6f5-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-kgxlg\" (UID: \"85b94340-ea1f-4cde-b359-d4970084e6f5\") " pod="openstack/dnsmasq-dns-845d6d6f59-kgxlg" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.193518 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.193588 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/85b94340-ea1f-4cde-b359-d4970084e6f5-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-kgxlg\" (UID: \"85b94340-ea1f-4cde-b359-d4970084e6f5\") " pod="openstack/dnsmasq-dns-845d6d6f59-kgxlg" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.193625 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzrc7\" (UniqueName: \"kubernetes.io/projected/0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d-kube-api-access-vzrc7\") pod \"nova-cell1-novncproxy-0\" (UID: \"0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.193659 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/85b94340-ea1f-4cde-b359-d4970084e6f5-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-kgxlg\" (UID: \"85b94340-ea1f-4cde-b359-d4970084e6f5\") " pod="openstack/dnsmasq-dns-845d6d6f59-kgxlg" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.193756 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.194625 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/85b94340-ea1f-4cde-b359-d4970084e6f5-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-kgxlg\" (UID: \"85b94340-ea1f-4cde-b359-d4970084e6f5\") " pod="openstack/dnsmasq-dns-845d6d6f59-kgxlg" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.195446 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85b94340-ea1f-4cde-b359-d4970084e6f5-config\") pod \"dnsmasq-dns-845d6d6f59-kgxlg\" (UID: \"85b94340-ea1f-4cde-b359-d4970084e6f5\") " pod="openstack/dnsmasq-dns-845d6d6f59-kgxlg" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.196461 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/85b94340-ea1f-4cde-b359-d4970084e6f5-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-kgxlg\" (UID: \"85b94340-ea1f-4cde-b359-d4970084e6f5\") " pod="openstack/dnsmasq-dns-845d6d6f59-kgxlg" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.206049 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.214632 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/85b94340-ea1f-4cde-b359-d4970084e6f5-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-kgxlg\" (UID: \"85b94340-ea1f-4cde-b359-d4970084e6f5\") " pod="openstack/dnsmasq-dns-845d6d6f59-kgxlg" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.214952 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/85b94340-ea1f-4cde-b359-d4970084e6f5-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-kgxlg\" (UID: \"85b94340-ea1f-4cde-b359-d4970084e6f5\") " pod="openstack/dnsmasq-dns-845d6d6f59-kgxlg" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.228658 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqxjp\" (UniqueName: \"kubernetes.io/projected/85b94340-ea1f-4cde-b359-d4970084e6f5-kube-api-access-rqxjp\") pod \"dnsmasq-dns-845d6d6f59-kgxlg\" (UID: \"85b94340-ea1f-4cde-b359-d4970084e6f5\") " pod="openstack/dnsmasq-dns-845d6d6f59-kgxlg" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.295841 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.296066 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzrc7\" (UniqueName: \"kubernetes.io/projected/0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d-kube-api-access-vzrc7\") pod \"nova-cell1-novncproxy-0\" (UID: \"0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.296341 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.303635 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.307230 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.313830 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzrc7\" (UniqueName: \"kubernetes.io/projected/0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d-kube-api-access-vzrc7\") pod \"nova-cell1-novncproxy-0\" (UID: \"0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.398371 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.424502 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-kgxlg" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.467527 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.571849 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-v9q8t"] Sep 30 14:15:08 crc kubenswrapper[4840]: W0930 14:15:08.601467 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod722796b2_e357_4fec_8437_46f163f510d6.slice/crio-3df6b7e93fd969828c9b208b9fa80502d897b379b7b65554d3825da4f48422bc WatchSource:0}: Error finding container 3df6b7e93fd969828c9b208b9fa80502d897b379b7b65554d3825da4f48422bc: Status 404 returned error can't find the container with id 3df6b7e93fd969828c9b208b9fa80502d897b379b7b65554d3825da4f48422bc Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.733319 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-v9q8t" event={"ID":"722796b2-e357-4fec-8437-46f163f510d6","Type":"ContainerStarted","Data":"3df6b7e93fd969828c9b208b9fa80502d897b379b7b65554d3825da4f48422bc"} Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.756880 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 14:15:08 crc kubenswrapper[4840]: W0930 14:15:08.767708 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod70d820fe_33fc_4d18_b75f_3bff2044ba5b.slice/crio-546e82255fc2365ea722d4b14726b297b66cfd14446d6bba819b50aee8b48a12 WatchSource:0}: Error finding container 546e82255fc2365ea722d4b14726b297b66cfd14446d6bba819b50aee8b48a12: Status 404 returned error can't find the container with id 546e82255fc2365ea722d4b14726b297b66cfd14446d6bba819b50aee8b48a12 Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.860811 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-2ph7k"] Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.862209 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-2ph7k" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.864526 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.865067 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Sep 30 14:15:08 crc kubenswrapper[4840]: W0930 14:15:08.872860 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc04dbcf_177f_4039_be37_9986232e2d42.slice/crio-7048570c998f9562d67033e22a65838b625dd1d6034b1449b6ce82318d28de6a WatchSource:0}: Error finding container 7048570c998f9562d67033e22a65838b625dd1d6034b1449b6ce82318d28de6a: Status 404 returned error can't find the container with id 7048570c998f9562d67033e22a65838b625dd1d6034b1449b6ce82318d28de6a Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.881306 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-2ph7k"] Sep 30 14:15:08 crc kubenswrapper[4840]: I0930 14:15:08.892156 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 30 14:15:09 crc kubenswrapper[4840]: I0930 14:15:09.013215 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91fc410c-6ef8-4864-8ba2-a604c7bcd8ac-scripts\") pod \"nova-cell1-conductor-db-sync-2ph7k\" (UID: \"91fc410c-6ef8-4864-8ba2-a604c7bcd8ac\") " pod="openstack/nova-cell1-conductor-db-sync-2ph7k" Sep 30 14:15:09 crc kubenswrapper[4840]: I0930 14:15:09.013348 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hkjq\" (UniqueName: \"kubernetes.io/projected/91fc410c-6ef8-4864-8ba2-a604c7bcd8ac-kube-api-access-2hkjq\") pod \"nova-cell1-conductor-db-sync-2ph7k\" (UID: \"91fc410c-6ef8-4864-8ba2-a604c7bcd8ac\") " pod="openstack/nova-cell1-conductor-db-sync-2ph7k" Sep 30 14:15:09 crc kubenswrapper[4840]: I0930 14:15:09.013379 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91fc410c-6ef8-4864-8ba2-a604c7bcd8ac-config-data\") pod \"nova-cell1-conductor-db-sync-2ph7k\" (UID: \"91fc410c-6ef8-4864-8ba2-a604c7bcd8ac\") " pod="openstack/nova-cell1-conductor-db-sync-2ph7k" Sep 30 14:15:09 crc kubenswrapper[4840]: I0930 14:15:09.013412 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91fc410c-6ef8-4864-8ba2-a604c7bcd8ac-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-2ph7k\" (UID: \"91fc410c-6ef8-4864-8ba2-a604c7bcd8ac\") " pod="openstack/nova-cell1-conductor-db-sync-2ph7k" Sep 30 14:15:09 crc kubenswrapper[4840]: I0930 14:15:09.025122 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 14:15:09 crc kubenswrapper[4840]: W0930 14:15:09.116445 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod85b94340_ea1f_4cde_b359_d4970084e6f5.slice/crio-12adbd12bf25d9bc20e7f291061be4abaa64109c9ce2809df8a0801723c1fe00 WatchSource:0}: Error finding container 12adbd12bf25d9bc20e7f291061be4abaa64109c9ce2809df8a0801723c1fe00: Status 404 returned error can't find the container with id 12adbd12bf25d9bc20e7f291061be4abaa64109c9ce2809df8a0801723c1fe00 Sep 30 14:15:09 crc kubenswrapper[4840]: I0930 14:15:09.117241 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91fc410c-6ef8-4864-8ba2-a604c7bcd8ac-scripts\") pod \"nova-cell1-conductor-db-sync-2ph7k\" (UID: \"91fc410c-6ef8-4864-8ba2-a604c7bcd8ac\") " pod="openstack/nova-cell1-conductor-db-sync-2ph7k" Sep 30 14:15:09 crc kubenswrapper[4840]: I0930 14:15:09.117464 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hkjq\" (UniqueName: \"kubernetes.io/projected/91fc410c-6ef8-4864-8ba2-a604c7bcd8ac-kube-api-access-2hkjq\") pod \"nova-cell1-conductor-db-sync-2ph7k\" (UID: \"91fc410c-6ef8-4864-8ba2-a604c7bcd8ac\") " pod="openstack/nova-cell1-conductor-db-sync-2ph7k" Sep 30 14:15:09 crc kubenswrapper[4840]: I0930 14:15:09.117510 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91fc410c-6ef8-4864-8ba2-a604c7bcd8ac-config-data\") pod \"nova-cell1-conductor-db-sync-2ph7k\" (UID: \"91fc410c-6ef8-4864-8ba2-a604c7bcd8ac\") " pod="openstack/nova-cell1-conductor-db-sync-2ph7k" Sep 30 14:15:09 crc kubenswrapper[4840]: I0930 14:15:09.117572 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91fc410c-6ef8-4864-8ba2-a604c7bcd8ac-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-2ph7k\" (UID: \"91fc410c-6ef8-4864-8ba2-a604c7bcd8ac\") " pod="openstack/nova-cell1-conductor-db-sync-2ph7k" Sep 30 14:15:09 crc kubenswrapper[4840]: I0930 14:15:09.118399 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-kgxlg"] Sep 30 14:15:09 crc kubenswrapper[4840]: I0930 14:15:09.126949 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91fc410c-6ef8-4864-8ba2-a604c7bcd8ac-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-2ph7k\" (UID: \"91fc410c-6ef8-4864-8ba2-a604c7bcd8ac\") " pod="openstack/nova-cell1-conductor-db-sync-2ph7k" Sep 30 14:15:09 crc kubenswrapper[4840]: I0930 14:15:09.127053 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91fc410c-6ef8-4864-8ba2-a604c7bcd8ac-config-data\") pod \"nova-cell1-conductor-db-sync-2ph7k\" (UID: \"91fc410c-6ef8-4864-8ba2-a604c7bcd8ac\") " pod="openstack/nova-cell1-conductor-db-sync-2ph7k" Sep 30 14:15:09 crc kubenswrapper[4840]: I0930 14:15:09.132787 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91fc410c-6ef8-4864-8ba2-a604c7bcd8ac-scripts\") pod \"nova-cell1-conductor-db-sync-2ph7k\" (UID: \"91fc410c-6ef8-4864-8ba2-a604c7bcd8ac\") " pod="openstack/nova-cell1-conductor-db-sync-2ph7k" Sep 30 14:15:09 crc kubenswrapper[4840]: I0930 14:15:09.136051 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hkjq\" (UniqueName: \"kubernetes.io/projected/91fc410c-6ef8-4864-8ba2-a604c7bcd8ac-kube-api-access-2hkjq\") pod \"nova-cell1-conductor-db-sync-2ph7k\" (UID: \"91fc410c-6ef8-4864-8ba2-a604c7bcd8ac\") " pod="openstack/nova-cell1-conductor-db-sync-2ph7k" Sep 30 14:15:09 crc kubenswrapper[4840]: I0930 14:15:09.197012 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 30 14:15:09 crc kubenswrapper[4840]: I0930 14:15:09.232924 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-2ph7k" Sep 30 14:15:09 crc kubenswrapper[4840]: I0930 14:15:09.698450 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-2ph7k"] Sep 30 14:15:09 crc kubenswrapper[4840]: I0930 14:15:09.774518 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-2ph7k" event={"ID":"91fc410c-6ef8-4864-8ba2-a604c7bcd8ac","Type":"ContainerStarted","Data":"2b27344500ffc8121defe005851d508edfe9d8afabbcbe19cec1a9a44ef6dbd7"} Sep 30 14:15:09 crc kubenswrapper[4840]: I0930 14:15:09.796535 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-v9q8t" event={"ID":"722796b2-e357-4fec-8437-46f163f510d6","Type":"ContainerStarted","Data":"8bdf30fd0b762dd4d4417dd4e2b2f0e0e131385af5e4454443d06237bb91c9be"} Sep 30 14:15:09 crc kubenswrapper[4840]: I0930 14:15:09.802624 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e616a00a-2b1e-4ea6-bde6-1a5517c8464e","Type":"ContainerStarted","Data":"2999182eef0c988cb9dec38883198a1f9d47f049d14417495090236013dfef1a"} Sep 30 14:15:09 crc kubenswrapper[4840]: I0930 14:15:09.809043 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d","Type":"ContainerStarted","Data":"91ac81520ec514e2569ca168378bf0052bc6172ba51c8488f0c4ccf92a2abfeb"} Sep 30 14:15:09 crc kubenswrapper[4840]: I0930 14:15:09.820080 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"70d820fe-33fc-4d18-b75f-3bff2044ba5b","Type":"ContainerStarted","Data":"546e82255fc2365ea722d4b14726b297b66cfd14446d6bba819b50aee8b48a12"} Sep 30 14:15:09 crc kubenswrapper[4840]: I0930 14:15:09.848313 4840 generic.go:334] "Generic (PLEG): container finished" podID="85b94340-ea1f-4cde-b359-d4970084e6f5" containerID="4fa5e7cb35fa61ade81028ac2dd0b73d7848ab07034be304329ea1bfb9d0bc25" exitCode=0 Sep 30 14:15:09 crc kubenswrapper[4840]: I0930 14:15:09.848432 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-kgxlg" event={"ID":"85b94340-ea1f-4cde-b359-d4970084e6f5","Type":"ContainerDied","Data":"4fa5e7cb35fa61ade81028ac2dd0b73d7848ab07034be304329ea1bfb9d0bc25"} Sep 30 14:15:09 crc kubenswrapper[4840]: I0930 14:15:09.848468 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-kgxlg" event={"ID":"85b94340-ea1f-4cde-b359-d4970084e6f5","Type":"ContainerStarted","Data":"12adbd12bf25d9bc20e7f291061be4abaa64109c9ce2809df8a0801723c1fe00"} Sep 30 14:15:09 crc kubenswrapper[4840]: I0930 14:15:09.852647 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-v9q8t" podStartSLOduration=2.852619736 podStartE2EDuration="2.852619736s" podCreationTimestamp="2025-09-30 14:15:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:15:09.817959892 +0000 UTC m=+1138.447046315" watchObservedRunningTime="2025-09-30 14:15:09.852619736 +0000 UTC m=+1138.481706159" Sep 30 14:15:09 crc kubenswrapper[4840]: I0930 14:15:09.858961 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"bc04dbcf-177f-4039-be37-9986232e2d42","Type":"ContainerStarted","Data":"7048570c998f9562d67033e22a65838b625dd1d6034b1449b6ce82318d28de6a"} Sep 30 14:15:10 crc kubenswrapper[4840]: I0930 14:15:10.870237 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-kgxlg" event={"ID":"85b94340-ea1f-4cde-b359-d4970084e6f5","Type":"ContainerStarted","Data":"9944d1240049ca763bd1030a91e7bfc104f71e03553fc6b95c95cf3a385e5784"} Sep 30 14:15:10 crc kubenswrapper[4840]: I0930 14:15:10.873002 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-845d6d6f59-kgxlg" Sep 30 14:15:10 crc kubenswrapper[4840]: I0930 14:15:10.876641 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-2ph7k" event={"ID":"91fc410c-6ef8-4864-8ba2-a604c7bcd8ac","Type":"ContainerStarted","Data":"d2a2e49e9d9bfd23804ae23d4fe6675ebfe92252517fe6c60b4fc629f086bff7"} Sep 30 14:15:10 crc kubenswrapper[4840]: I0930 14:15:10.897300 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-845d6d6f59-kgxlg" podStartSLOduration=3.897279344 podStartE2EDuration="3.897279344s" podCreationTimestamp="2025-09-30 14:15:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:15:10.893152116 +0000 UTC m=+1139.522238559" watchObservedRunningTime="2025-09-30 14:15:10.897279344 +0000 UTC m=+1139.526365767" Sep 30 14:15:10 crc kubenswrapper[4840]: I0930 14:15:10.911128 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-2ph7k" podStartSLOduration=2.911108914 podStartE2EDuration="2.911108914s" podCreationTimestamp="2025-09-30 14:15:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:15:10.907214623 +0000 UTC m=+1139.536301046" watchObservedRunningTime="2025-09-30 14:15:10.911108914 +0000 UTC m=+1139.540195337" Sep 30 14:15:11 crc kubenswrapper[4840]: I0930 14:15:11.719467 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 30 14:15:11 crc kubenswrapper[4840]: I0930 14:15:11.832392 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 14:15:12 crc kubenswrapper[4840]: I0930 14:15:12.836507 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Sep 30 14:15:15 crc kubenswrapper[4840]: I0930 14:15:15.929135 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"bc04dbcf-177f-4039-be37-9986232e2d42","Type":"ContainerStarted","Data":"2c0a5ef199487715a5f0c8297b85a8e37514dd7d080cbaa9715e7083c1447fa7"} Sep 30 14:15:15 crc kubenswrapper[4840]: I0930 14:15:15.930893 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e616a00a-2b1e-4ea6-bde6-1a5517c8464e","Type":"ContainerStarted","Data":"d0a85ca9716a373dcf69f3c04d2251d1426d5a6fadaaf28100c017ceffd8affb"} Sep 30 14:15:15 crc kubenswrapper[4840]: I0930 14:15:15.934088 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"70d820fe-33fc-4d18-b75f-3bff2044ba5b","Type":"ContainerStarted","Data":"fac2744bf53694103c7828bf5f0188b8f2b6fedbd8a1401d6de5794617e09f68"} Sep 30 14:15:16 crc kubenswrapper[4840]: I0930 14:15:16.713435 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 30 14:15:16 crc kubenswrapper[4840]: I0930 14:15:16.716294 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="3e51f892-d5b7-48b6-9a5f-0c0af8a00f21" containerName="kube-state-metrics" containerID="cri-o://a8c84f922f39612d031013b9885fdcb52dcc7faafb7a80671ecfc2c1353cc4e6" gracePeriod=30 Sep 30 14:15:16 crc kubenswrapper[4840]: I0930 14:15:16.946698 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d","Type":"ContainerStarted","Data":"8a34fb015dd622d5f4779489064636d959dba820baf62e3f5f3b9216bbcc2214"} Sep 30 14:15:16 crc kubenswrapper[4840]: I0930 14:15:16.946759 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://8a34fb015dd622d5f4779489064636d959dba820baf62e3f5f3b9216bbcc2214" gracePeriod=30 Sep 30 14:15:16 crc kubenswrapper[4840]: I0930 14:15:16.952126 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"70d820fe-33fc-4d18-b75f-3bff2044ba5b","Type":"ContainerStarted","Data":"b99bb2a6e98a29ed72b883c750086a4337e5a42a0246cb6616a4a480e7e68f76"} Sep 30 14:15:16 crc kubenswrapper[4840]: I0930 14:15:16.954941 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"bc04dbcf-177f-4039-be37-9986232e2d42","Type":"ContainerStarted","Data":"d0e85f6c23c257ef6bbd38b05fea7e9d3832058240ab6ae45ee6c8d982f17c47"} Sep 30 14:15:16 crc kubenswrapper[4840]: I0930 14:15:16.967362 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.592316765 podStartE2EDuration="9.967336713s" podCreationTimestamp="2025-09-30 14:15:07 +0000 UTC" firstStartedPulling="2025-09-30 14:15:09.209193595 +0000 UTC m=+1137.838280018" lastFinishedPulling="2025-09-30 14:15:15.584213543 +0000 UTC m=+1144.213299966" observedRunningTime="2025-09-30 14:15:16.964021417 +0000 UTC m=+1145.593107850" watchObservedRunningTime="2025-09-30 14:15:16.967336713 +0000 UTC m=+1145.596423136" Sep 30 14:15:16 crc kubenswrapper[4840]: I0930 14:15:16.992133 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.437287785 podStartE2EDuration="9.992110719s" podCreationTimestamp="2025-09-30 14:15:07 +0000 UTC" firstStartedPulling="2025-09-30 14:15:09.029229365 +0000 UTC m=+1137.658315788" lastFinishedPulling="2025-09-30 14:15:15.584052309 +0000 UTC m=+1144.213138722" observedRunningTime="2025-09-30 14:15:16.982666033 +0000 UTC m=+1145.611752456" watchObservedRunningTime="2025-09-30 14:15:16.992110719 +0000 UTC m=+1145.621197152" Sep 30 14:15:17 crc kubenswrapper[4840]: I0930 14:15:17.964636 4840 generic.go:334] "Generic (PLEG): container finished" podID="722796b2-e357-4fec-8437-46f163f510d6" containerID="8bdf30fd0b762dd4d4417dd4e2b2f0e0e131385af5e4454443d06237bb91c9be" exitCode=0 Sep 30 14:15:17 crc kubenswrapper[4840]: I0930 14:15:17.964705 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-v9q8t" event={"ID":"722796b2-e357-4fec-8437-46f163f510d6","Type":"ContainerDied","Data":"8bdf30fd0b762dd4d4417dd4e2b2f0e0e131385af5e4454443d06237bb91c9be"} Sep 30 14:15:17 crc kubenswrapper[4840]: I0930 14:15:17.966371 4840 generic.go:334] "Generic (PLEG): container finished" podID="3e51f892-d5b7-48b6-9a5f-0c0af8a00f21" containerID="a8c84f922f39612d031013b9885fdcb52dcc7faafb7a80671ecfc2c1353cc4e6" exitCode=2 Sep 30 14:15:17 crc kubenswrapper[4840]: I0930 14:15:17.966435 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3e51f892-d5b7-48b6-9a5f-0c0af8a00f21","Type":"ContainerDied","Data":"a8c84f922f39612d031013b9885fdcb52dcc7faafb7a80671ecfc2c1353cc4e6"} Sep 30 14:15:17 crc kubenswrapper[4840]: I0930 14:15:17.966593 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="70d820fe-33fc-4d18-b75f-3bff2044ba5b" containerName="nova-metadata-metadata" containerID="cri-o://b99bb2a6e98a29ed72b883c750086a4337e5a42a0246cb6616a4a480e7e68f76" gracePeriod=30 Sep 30 14:15:17 crc kubenswrapper[4840]: I0930 14:15:17.966610 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="70d820fe-33fc-4d18-b75f-3bff2044ba5b" containerName="nova-metadata-log" containerID="cri-o://fac2744bf53694103c7828bf5f0188b8f2b6fedbd8a1401d6de5794617e09f68" gracePeriod=30 Sep 30 14:15:17 crc kubenswrapper[4840]: I0930 14:15:17.999574 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=4.287286258 podStartE2EDuration="10.999538296s" podCreationTimestamp="2025-09-30 14:15:07 +0000 UTC" firstStartedPulling="2025-09-30 14:15:08.878392323 +0000 UTC m=+1137.507478746" lastFinishedPulling="2025-09-30 14:15:15.590644361 +0000 UTC m=+1144.219730784" observedRunningTime="2025-09-30 14:15:17.993754566 +0000 UTC m=+1146.622840989" watchObservedRunningTime="2025-09-30 14:15:17.999538296 +0000 UTC m=+1146.628624729" Sep 30 14:15:18 crc kubenswrapper[4840]: I0930 14:15:18.015578 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=4.20753059 podStartE2EDuration="11.015562254s" podCreationTimestamp="2025-09-30 14:15:07 +0000 UTC" firstStartedPulling="2025-09-30 14:15:08.777204116 +0000 UTC m=+1137.406290539" lastFinishedPulling="2025-09-30 14:15:15.58523579 +0000 UTC m=+1144.214322203" observedRunningTime="2025-09-30 14:15:18.013707136 +0000 UTC m=+1146.642793569" watchObservedRunningTime="2025-09-30 14:15:18.015562254 +0000 UTC m=+1146.644648677" Sep 30 14:15:18 crc kubenswrapper[4840]: I0930 14:15:18.171335 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 30 14:15:18 crc kubenswrapper[4840]: I0930 14:15:18.171689 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 30 14:15:18 crc kubenswrapper[4840]: I0930 14:15:18.206715 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 30 14:15:18 crc kubenswrapper[4840]: I0930 14:15:18.206758 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 30 14:15:18 crc kubenswrapper[4840]: I0930 14:15:18.399280 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Sep 30 14:15:18 crc kubenswrapper[4840]: I0930 14:15:18.399588 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Sep 30 14:15:18 crc kubenswrapper[4840]: I0930 14:15:18.406683 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 30 14:15:18 crc kubenswrapper[4840]: I0930 14:15:18.426703 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-845d6d6f59-kgxlg" Sep 30 14:15:18 crc kubenswrapper[4840]: I0930 14:15:18.434041 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Sep 30 14:15:18 crc kubenswrapper[4840]: I0930 14:15:18.467723 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:15:18 crc kubenswrapper[4840]: I0930 14:15:18.513473 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-qs9s6"] Sep 30 14:15:18 crc kubenswrapper[4840]: I0930 14:15:18.513726 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5784cf869f-qs9s6" podUID="ae2f3329-33ca-402d-b841-8f1e257bbf8a" containerName="dnsmasq-dns" containerID="cri-o://71b5753eee0c71364cb76a4843975d6ba07a09e3c6338c5ac66b32b4a82d65d0" gracePeriod=10 Sep 30 14:15:18 crc kubenswrapper[4840]: I0930 14:15:18.516055 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7mxd\" (UniqueName: \"kubernetes.io/projected/3e51f892-d5b7-48b6-9a5f-0c0af8a00f21-kube-api-access-w7mxd\") pod \"3e51f892-d5b7-48b6-9a5f-0c0af8a00f21\" (UID: \"3e51f892-d5b7-48b6-9a5f-0c0af8a00f21\") " Sep 30 14:15:18 crc kubenswrapper[4840]: I0930 14:15:18.524279 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e51f892-d5b7-48b6-9a5f-0c0af8a00f21-kube-api-access-w7mxd" (OuterVolumeSpecName: "kube-api-access-w7mxd") pod "3e51f892-d5b7-48b6-9a5f-0c0af8a00f21" (UID: "3e51f892-d5b7-48b6-9a5f-0c0af8a00f21"). InnerVolumeSpecName "kube-api-access-w7mxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:15:18 crc kubenswrapper[4840]: I0930 14:15:18.618164 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7mxd\" (UniqueName: \"kubernetes.io/projected/3e51f892-d5b7-48b6-9a5f-0c0af8a00f21-kube-api-access-w7mxd\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:18 crc kubenswrapper[4840]: I0930 14:15:18.976075 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3e51f892-d5b7-48b6-9a5f-0c0af8a00f21","Type":"ContainerDied","Data":"46e2333b7edcfa1997bd02cd50e7f99107645217b176097071a260c0a547eeb2"} Sep 30 14:15:18 crc kubenswrapper[4840]: I0930 14:15:18.976426 4840 scope.go:117] "RemoveContainer" containerID="a8c84f922f39612d031013b9885fdcb52dcc7faafb7a80671ecfc2c1353cc4e6" Sep 30 14:15:18 crc kubenswrapper[4840]: I0930 14:15:18.976089 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 30 14:15:18 crc kubenswrapper[4840]: I0930 14:15:18.979935 4840 generic.go:334] "Generic (PLEG): container finished" podID="70d820fe-33fc-4d18-b75f-3bff2044ba5b" containerID="b99bb2a6e98a29ed72b883c750086a4337e5a42a0246cb6616a4a480e7e68f76" exitCode=0 Sep 30 14:15:18 crc kubenswrapper[4840]: I0930 14:15:18.979961 4840 generic.go:334] "Generic (PLEG): container finished" podID="70d820fe-33fc-4d18-b75f-3bff2044ba5b" containerID="fac2744bf53694103c7828bf5f0188b8f2b6fedbd8a1401d6de5794617e09f68" exitCode=143 Sep 30 14:15:18 crc kubenswrapper[4840]: I0930 14:15:18.979964 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"70d820fe-33fc-4d18-b75f-3bff2044ba5b","Type":"ContainerDied","Data":"b99bb2a6e98a29ed72b883c750086a4337e5a42a0246cb6616a4a480e7e68f76"} Sep 30 14:15:18 crc kubenswrapper[4840]: I0930 14:15:18.979992 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"70d820fe-33fc-4d18-b75f-3bff2044ba5b","Type":"ContainerDied","Data":"fac2744bf53694103c7828bf5f0188b8f2b6fedbd8a1401d6de5794617e09f68"} Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.012172 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.018731 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.037606 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.048483 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Sep 30 14:15:19 crc kubenswrapper[4840]: E0930 14:15:19.048875 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e51f892-d5b7-48b6-9a5f-0c0af8a00f21" containerName="kube-state-metrics" Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.048891 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e51f892-d5b7-48b6-9a5f-0c0af8a00f21" containerName="kube-state-metrics" Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.049107 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e51f892-d5b7-48b6-9a5f-0c0af8a00f21" containerName="kube-state-metrics" Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.049795 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.065904 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.065980 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.080851 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.127301 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/26d4f9e8-9d18-45ec-8117-39c0714ce1b0-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"26d4f9e8-9d18-45ec-8117-39c0714ce1b0\") " pod="openstack/kube-state-metrics-0" Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.139152 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26d4f9e8-9d18-45ec-8117-39c0714ce1b0-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"26d4f9e8-9d18-45ec-8117-39c0714ce1b0\") " pod="openstack/kube-state-metrics-0" Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.139289 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvdtg\" (UniqueName: \"kubernetes.io/projected/26d4f9e8-9d18-45ec-8117-39c0714ce1b0-kube-api-access-gvdtg\") pod \"kube-state-metrics-0\" (UID: \"26d4f9e8-9d18-45ec-8117-39c0714ce1b0\") " pod="openstack/kube-state-metrics-0" Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.139370 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/26d4f9e8-9d18-45ec-8117-39c0714ce1b0-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"26d4f9e8-9d18-45ec-8117-39c0714ce1b0\") " pod="openstack/kube-state-metrics-0" Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.241320 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/26d4f9e8-9d18-45ec-8117-39c0714ce1b0-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"26d4f9e8-9d18-45ec-8117-39c0714ce1b0\") " pod="openstack/kube-state-metrics-0" Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.241458 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/26d4f9e8-9d18-45ec-8117-39c0714ce1b0-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"26d4f9e8-9d18-45ec-8117-39c0714ce1b0\") " pod="openstack/kube-state-metrics-0" Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.241687 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26d4f9e8-9d18-45ec-8117-39c0714ce1b0-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"26d4f9e8-9d18-45ec-8117-39c0714ce1b0\") " pod="openstack/kube-state-metrics-0" Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.241715 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvdtg\" (UniqueName: \"kubernetes.io/projected/26d4f9e8-9d18-45ec-8117-39c0714ce1b0-kube-api-access-gvdtg\") pod \"kube-state-metrics-0\" (UID: \"26d4f9e8-9d18-45ec-8117-39c0714ce1b0\") " pod="openstack/kube-state-metrics-0" Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.246859 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/26d4f9e8-9d18-45ec-8117-39c0714ce1b0-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"26d4f9e8-9d18-45ec-8117-39c0714ce1b0\") " pod="openstack/kube-state-metrics-0" Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.247221 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26d4f9e8-9d18-45ec-8117-39c0714ce1b0-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"26d4f9e8-9d18-45ec-8117-39c0714ce1b0\") " pod="openstack/kube-state-metrics-0" Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.251355 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/26d4f9e8-9d18-45ec-8117-39c0714ce1b0-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"26d4f9e8-9d18-45ec-8117-39c0714ce1b0\") " pod="openstack/kube-state-metrics-0" Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.258837 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvdtg\" (UniqueName: \"kubernetes.io/projected/26d4f9e8-9d18-45ec-8117-39c0714ce1b0-kube-api-access-gvdtg\") pod \"kube-state-metrics-0\" (UID: \"26d4f9e8-9d18-45ec-8117-39c0714ce1b0\") " pod="openstack/kube-state-metrics-0" Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.290923 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="bc04dbcf-177f-4039-be37-9986232e2d42" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.188:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.291319 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="bc04dbcf-177f-4039-be37-9986232e2d42" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.188:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.411848 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-v9q8t" Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.421532 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.551309 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/722796b2-e357-4fec-8437-46f163f510d6-config-data\") pod \"722796b2-e357-4fec-8437-46f163f510d6\" (UID: \"722796b2-e357-4fec-8437-46f163f510d6\") " Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.551407 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/722796b2-e357-4fec-8437-46f163f510d6-combined-ca-bundle\") pod \"722796b2-e357-4fec-8437-46f163f510d6\" (UID: \"722796b2-e357-4fec-8437-46f163f510d6\") " Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.551426 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/722796b2-e357-4fec-8437-46f163f510d6-scripts\") pod \"722796b2-e357-4fec-8437-46f163f510d6\" (UID: \"722796b2-e357-4fec-8437-46f163f510d6\") " Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.551515 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxrxr\" (UniqueName: \"kubernetes.io/projected/722796b2-e357-4fec-8437-46f163f510d6-kube-api-access-mxrxr\") pod \"722796b2-e357-4fec-8437-46f163f510d6\" (UID: \"722796b2-e357-4fec-8437-46f163f510d6\") " Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.557059 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/722796b2-e357-4fec-8437-46f163f510d6-kube-api-access-mxrxr" (OuterVolumeSpecName: "kube-api-access-mxrxr") pod "722796b2-e357-4fec-8437-46f163f510d6" (UID: "722796b2-e357-4fec-8437-46f163f510d6"). InnerVolumeSpecName "kube-api-access-mxrxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.582409 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/722796b2-e357-4fec-8437-46f163f510d6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "722796b2-e357-4fec-8437-46f163f510d6" (UID: "722796b2-e357-4fec-8437-46f163f510d6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.597490 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/722796b2-e357-4fec-8437-46f163f510d6-config-data" (OuterVolumeSpecName: "config-data") pod "722796b2-e357-4fec-8437-46f163f510d6" (UID: "722796b2-e357-4fec-8437-46f163f510d6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.654224 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxrxr\" (UniqueName: \"kubernetes.io/projected/722796b2-e357-4fec-8437-46f163f510d6-kube-api-access-mxrxr\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.654250 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/722796b2-e357-4fec-8437-46f163f510d6-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.654259 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/722796b2-e357-4fec-8437-46f163f510d6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.664721 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/722796b2-e357-4fec-8437-46f163f510d6-scripts" (OuterVolumeSpecName: "scripts") pod "722796b2-e357-4fec-8437-46f163f510d6" (UID: "722796b2-e357-4fec-8437-46f163f510d6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.755781 4840 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/722796b2-e357-4fec-8437-46f163f510d6-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.845197 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.846045 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="712feb06-cb97-425d-8b75-fcd5815a94ab" containerName="proxy-httpd" containerID="cri-o://25e57148f578d9ab12ef79020c7ac3c16f6b8542b50b2833a6d175c2d8a4c3f0" gracePeriod=30 Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.846178 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="712feb06-cb97-425d-8b75-fcd5815a94ab" containerName="sg-core" containerID="cri-o://53aa98fbc16bc77ba4f16d0623177aa092f9f3f0a29df3651490494fd3510c5d" gracePeriod=30 Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.846237 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="712feb06-cb97-425d-8b75-fcd5815a94ab" containerName="ceilometer-notification-agent" containerID="cri-o://fa186597d4025b0b4a5b61ec12ce99d1733596a1825f242ee2bcbfc59b2000c6" gracePeriod=30 Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.845469 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="712feb06-cb97-425d-8b75-fcd5815a94ab" containerName="ceilometer-central-agent" containerID="cri-o://9ffdf9131bd8af23b5cc76199769f964619ded52d0516748629ec0e503506d3c" gracePeriod=30 Sep 30 14:15:19 crc kubenswrapper[4840]: I0930 14:15:19.891184 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 30 14:15:19 crc kubenswrapper[4840]: W0930 14:15:19.903302 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod26d4f9e8_9d18_45ec_8117_39c0714ce1b0.slice/crio-e191deaa7860609744e1397fff665e7348cec71ebbac0a29879532ac51ceca25 WatchSource:0}: Error finding container e191deaa7860609744e1397fff665e7348cec71ebbac0a29879532ac51ceca25: Status 404 returned error can't find the container with id e191deaa7860609744e1397fff665e7348cec71ebbac0a29879532ac51ceca25 Sep 30 14:15:20 crc kubenswrapper[4840]: I0930 14:15:20.016115 4840 generic.go:334] "Generic (PLEG): container finished" podID="ae2f3329-33ca-402d-b841-8f1e257bbf8a" containerID="71b5753eee0c71364cb76a4843975d6ba07a09e3c6338c5ac66b32b4a82d65d0" exitCode=0 Sep 30 14:15:20 crc kubenswrapper[4840]: I0930 14:15:20.016227 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-qs9s6" event={"ID":"ae2f3329-33ca-402d-b841-8f1e257bbf8a","Type":"ContainerDied","Data":"71b5753eee0c71364cb76a4843975d6ba07a09e3c6338c5ac66b32b4a82d65d0"} Sep 30 14:15:20 crc kubenswrapper[4840]: I0930 14:15:20.019463 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-v9q8t" event={"ID":"722796b2-e357-4fec-8437-46f163f510d6","Type":"ContainerDied","Data":"3df6b7e93fd969828c9b208b9fa80502d897b379b7b65554d3825da4f48422bc"} Sep 30 14:15:20 crc kubenswrapper[4840]: I0930 14:15:20.019495 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-v9q8t" Sep 30 14:15:20 crc kubenswrapper[4840]: I0930 14:15:20.019544 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3df6b7e93fd969828c9b208b9fa80502d897b379b7b65554d3825da4f48422bc" Sep 30 14:15:20 crc kubenswrapper[4840]: I0930 14:15:20.029631 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"26d4f9e8-9d18-45ec-8117-39c0714ce1b0","Type":"ContainerStarted","Data":"e191deaa7860609744e1397fff665e7348cec71ebbac0a29879532ac51ceca25"} Sep 30 14:15:20 crc kubenswrapper[4840]: I0930 14:15:20.045215 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 14:15:20 crc kubenswrapper[4840]: I0930 14:15:20.138400 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e51f892-d5b7-48b6-9a5f-0c0af8a00f21" path="/var/lib/kubelet/pods/3e51f892-d5b7-48b6-9a5f-0c0af8a00f21/volumes" Sep 30 14:15:20 crc kubenswrapper[4840]: I0930 14:15:20.156850 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 30 14:15:20 crc kubenswrapper[4840]: I0930 14:15:20.157062 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="bc04dbcf-177f-4039-be37-9986232e2d42" containerName="nova-api-log" containerID="cri-o://2c0a5ef199487715a5f0c8297b85a8e37514dd7d080cbaa9715e7083c1447fa7" gracePeriod=30 Sep 30 14:15:20 crc kubenswrapper[4840]: I0930 14:15:20.157512 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="bc04dbcf-177f-4039-be37-9986232e2d42" containerName="nova-api-api" containerID="cri-o://d0e85f6c23c257ef6bbd38b05fea7e9d3832058240ab6ae45ee6c8d982f17c47" gracePeriod=30 Sep 30 14:15:20 crc kubenswrapper[4840]: I0930 14:15:20.169708 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70d820fe-33fc-4d18-b75f-3bff2044ba5b-logs\") pod \"70d820fe-33fc-4d18-b75f-3bff2044ba5b\" (UID: \"70d820fe-33fc-4d18-b75f-3bff2044ba5b\") " Sep 30 14:15:20 crc kubenswrapper[4840]: I0930 14:15:20.169776 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fw4kk\" (UniqueName: \"kubernetes.io/projected/70d820fe-33fc-4d18-b75f-3bff2044ba5b-kube-api-access-fw4kk\") pod \"70d820fe-33fc-4d18-b75f-3bff2044ba5b\" (UID: \"70d820fe-33fc-4d18-b75f-3bff2044ba5b\") " Sep 30 14:15:20 crc kubenswrapper[4840]: I0930 14:15:20.169833 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70d820fe-33fc-4d18-b75f-3bff2044ba5b-combined-ca-bundle\") pod \"70d820fe-33fc-4d18-b75f-3bff2044ba5b\" (UID: \"70d820fe-33fc-4d18-b75f-3bff2044ba5b\") " Sep 30 14:15:20 crc kubenswrapper[4840]: I0930 14:15:20.169960 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70d820fe-33fc-4d18-b75f-3bff2044ba5b-config-data\") pod \"70d820fe-33fc-4d18-b75f-3bff2044ba5b\" (UID: \"70d820fe-33fc-4d18-b75f-3bff2044ba5b\") " Sep 30 14:15:20 crc kubenswrapper[4840]: I0930 14:15:20.170815 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70d820fe-33fc-4d18-b75f-3bff2044ba5b-logs" (OuterVolumeSpecName: "logs") pod "70d820fe-33fc-4d18-b75f-3bff2044ba5b" (UID: "70d820fe-33fc-4d18-b75f-3bff2044ba5b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:15:20 crc kubenswrapper[4840]: I0930 14:15:20.188845 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70d820fe-33fc-4d18-b75f-3bff2044ba5b-kube-api-access-fw4kk" (OuterVolumeSpecName: "kube-api-access-fw4kk") pod "70d820fe-33fc-4d18-b75f-3bff2044ba5b" (UID: "70d820fe-33fc-4d18-b75f-3bff2044ba5b"). InnerVolumeSpecName "kube-api-access-fw4kk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:15:20 crc kubenswrapper[4840]: I0930 14:15:20.235180 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70d820fe-33fc-4d18-b75f-3bff2044ba5b-config-data" (OuterVolumeSpecName: "config-data") pod "70d820fe-33fc-4d18-b75f-3bff2044ba5b" (UID: "70d820fe-33fc-4d18-b75f-3bff2044ba5b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:15:20 crc kubenswrapper[4840]: I0930 14:15:20.235440 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70d820fe-33fc-4d18-b75f-3bff2044ba5b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "70d820fe-33fc-4d18-b75f-3bff2044ba5b" (UID: "70d820fe-33fc-4d18-b75f-3bff2044ba5b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:15:20 crc kubenswrapper[4840]: I0930 14:15:20.275688 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fw4kk\" (UniqueName: \"kubernetes.io/projected/70d820fe-33fc-4d18-b75f-3bff2044ba5b-kube-api-access-fw4kk\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:20 crc kubenswrapper[4840]: I0930 14:15:20.275733 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70d820fe-33fc-4d18-b75f-3bff2044ba5b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:20 crc kubenswrapper[4840]: I0930 14:15:20.275742 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70d820fe-33fc-4d18-b75f-3bff2044ba5b-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:20 crc kubenswrapper[4840]: I0930 14:15:20.275751 4840 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70d820fe-33fc-4d18-b75f-3bff2044ba5b-logs\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:20 crc kubenswrapper[4840]: I0930 14:15:20.443015 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.042300 4840 generic.go:334] "Generic (PLEG): container finished" podID="bc04dbcf-177f-4039-be37-9986232e2d42" containerID="2c0a5ef199487715a5f0c8297b85a8e37514dd7d080cbaa9715e7083c1447fa7" exitCode=143 Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.042491 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"bc04dbcf-177f-4039-be37-9986232e2d42","Type":"ContainerDied","Data":"2c0a5ef199487715a5f0c8297b85a8e37514dd7d080cbaa9715e7083c1447fa7"} Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.045907 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"70d820fe-33fc-4d18-b75f-3bff2044ba5b","Type":"ContainerDied","Data":"546e82255fc2365ea722d4b14726b297b66cfd14446d6bba819b50aee8b48a12"} Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.046090 4840 scope.go:117] "RemoveContainer" containerID="b99bb2a6e98a29ed72b883c750086a4337e5a42a0246cb6616a4a480e7e68f76" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.046063 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.051406 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"712feb06-cb97-425d-8b75-fcd5815a94ab","Type":"ContainerDied","Data":"25e57148f578d9ab12ef79020c7ac3c16f6b8542b50b2833a6d175c2d8a4c3f0"} Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.052438 4840 generic.go:334] "Generic (PLEG): container finished" podID="712feb06-cb97-425d-8b75-fcd5815a94ab" containerID="25e57148f578d9ab12ef79020c7ac3c16f6b8542b50b2833a6d175c2d8a4c3f0" exitCode=0 Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.052604 4840 generic.go:334] "Generic (PLEG): container finished" podID="712feb06-cb97-425d-8b75-fcd5815a94ab" containerID="53aa98fbc16bc77ba4f16d0623177aa092f9f3f0a29df3651490494fd3510c5d" exitCode=2 Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.052692 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"712feb06-cb97-425d-8b75-fcd5815a94ab","Type":"ContainerDied","Data":"53aa98fbc16bc77ba4f16d0623177aa092f9f3f0a29df3651490494fd3510c5d"} Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.075988 4840 scope.go:117] "RemoveContainer" containerID="fac2744bf53694103c7828bf5f0188b8f2b6fedbd8a1401d6de5794617e09f68" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.092866 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.132205 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.149258 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Sep 30 14:15:21 crc kubenswrapper[4840]: E0930 14:15:21.150015 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70d820fe-33fc-4d18-b75f-3bff2044ba5b" containerName="nova-metadata-log" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.150031 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="70d820fe-33fc-4d18-b75f-3bff2044ba5b" containerName="nova-metadata-log" Sep 30 14:15:21 crc kubenswrapper[4840]: E0930 14:15:21.150045 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70d820fe-33fc-4d18-b75f-3bff2044ba5b" containerName="nova-metadata-metadata" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.150051 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="70d820fe-33fc-4d18-b75f-3bff2044ba5b" containerName="nova-metadata-metadata" Sep 30 14:15:21 crc kubenswrapper[4840]: E0930 14:15:21.150071 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="722796b2-e357-4fec-8437-46f163f510d6" containerName="nova-manage" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.150078 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="722796b2-e357-4fec-8437-46f163f510d6" containerName="nova-manage" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.150264 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="70d820fe-33fc-4d18-b75f-3bff2044ba5b" containerName="nova-metadata-log" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.150276 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="70d820fe-33fc-4d18-b75f-3bff2044ba5b" containerName="nova-metadata-metadata" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.150296 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="722796b2-e357-4fec-8437-46f163f510d6" containerName="nova-manage" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.151764 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.156284 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.157418 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.177590 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.292731 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a14c8d91-8aff-4a17-8c9f-ce3839b3dbed-logs\") pod \"nova-metadata-0\" (UID: \"a14c8d91-8aff-4a17-8c9f-ce3839b3dbed\") " pod="openstack/nova-metadata-0" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.293176 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a14c8d91-8aff-4a17-8c9f-ce3839b3dbed-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a14c8d91-8aff-4a17-8c9f-ce3839b3dbed\") " pod="openstack/nova-metadata-0" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.293342 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzf6n\" (UniqueName: \"kubernetes.io/projected/a14c8d91-8aff-4a17-8c9f-ce3839b3dbed-kube-api-access-jzf6n\") pod \"nova-metadata-0\" (UID: \"a14c8d91-8aff-4a17-8c9f-ce3839b3dbed\") " pod="openstack/nova-metadata-0" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.293430 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a14c8d91-8aff-4a17-8c9f-ce3839b3dbed-config-data\") pod \"nova-metadata-0\" (UID: \"a14c8d91-8aff-4a17-8c9f-ce3839b3dbed\") " pod="openstack/nova-metadata-0" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.293522 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a14c8d91-8aff-4a17-8c9f-ce3839b3dbed-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a14c8d91-8aff-4a17-8c9f-ce3839b3dbed\") " pod="openstack/nova-metadata-0" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.395570 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a14c8d91-8aff-4a17-8c9f-ce3839b3dbed-config-data\") pod \"nova-metadata-0\" (UID: \"a14c8d91-8aff-4a17-8c9f-ce3839b3dbed\") " pod="openstack/nova-metadata-0" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.395651 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a14c8d91-8aff-4a17-8c9f-ce3839b3dbed-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a14c8d91-8aff-4a17-8c9f-ce3839b3dbed\") " pod="openstack/nova-metadata-0" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.395737 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a14c8d91-8aff-4a17-8c9f-ce3839b3dbed-logs\") pod \"nova-metadata-0\" (UID: \"a14c8d91-8aff-4a17-8c9f-ce3839b3dbed\") " pod="openstack/nova-metadata-0" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.395798 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a14c8d91-8aff-4a17-8c9f-ce3839b3dbed-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a14c8d91-8aff-4a17-8c9f-ce3839b3dbed\") " pod="openstack/nova-metadata-0" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.395886 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzf6n\" (UniqueName: \"kubernetes.io/projected/a14c8d91-8aff-4a17-8c9f-ce3839b3dbed-kube-api-access-jzf6n\") pod \"nova-metadata-0\" (UID: \"a14c8d91-8aff-4a17-8c9f-ce3839b3dbed\") " pod="openstack/nova-metadata-0" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.397053 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a14c8d91-8aff-4a17-8c9f-ce3839b3dbed-logs\") pod \"nova-metadata-0\" (UID: \"a14c8d91-8aff-4a17-8c9f-ce3839b3dbed\") " pod="openstack/nova-metadata-0" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.401870 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a14c8d91-8aff-4a17-8c9f-ce3839b3dbed-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a14c8d91-8aff-4a17-8c9f-ce3839b3dbed\") " pod="openstack/nova-metadata-0" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.401940 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a14c8d91-8aff-4a17-8c9f-ce3839b3dbed-config-data\") pod \"nova-metadata-0\" (UID: \"a14c8d91-8aff-4a17-8c9f-ce3839b3dbed\") " pod="openstack/nova-metadata-0" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.402083 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a14c8d91-8aff-4a17-8c9f-ce3839b3dbed-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a14c8d91-8aff-4a17-8c9f-ce3839b3dbed\") " pod="openstack/nova-metadata-0" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.413169 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzf6n\" (UniqueName: \"kubernetes.io/projected/a14c8d91-8aff-4a17-8c9f-ce3839b3dbed-kube-api-access-jzf6n\") pod \"nova-metadata-0\" (UID: \"a14c8d91-8aff-4a17-8c9f-ce3839b3dbed\") " pod="openstack/nova-metadata-0" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.477442 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.498432 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-qs9s6" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.598864 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ae2f3329-33ca-402d-b841-8f1e257bbf8a-ovsdbserver-sb\") pod \"ae2f3329-33ca-402d-b841-8f1e257bbf8a\" (UID: \"ae2f3329-33ca-402d-b841-8f1e257bbf8a\") " Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.599225 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae2f3329-33ca-402d-b841-8f1e257bbf8a-config\") pod \"ae2f3329-33ca-402d-b841-8f1e257bbf8a\" (UID: \"ae2f3329-33ca-402d-b841-8f1e257bbf8a\") " Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.599317 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrvf9\" (UniqueName: \"kubernetes.io/projected/ae2f3329-33ca-402d-b841-8f1e257bbf8a-kube-api-access-nrvf9\") pod \"ae2f3329-33ca-402d-b841-8f1e257bbf8a\" (UID: \"ae2f3329-33ca-402d-b841-8f1e257bbf8a\") " Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.599349 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae2f3329-33ca-402d-b841-8f1e257bbf8a-dns-svc\") pod \"ae2f3329-33ca-402d-b841-8f1e257bbf8a\" (UID: \"ae2f3329-33ca-402d-b841-8f1e257bbf8a\") " Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.599391 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae2f3329-33ca-402d-b841-8f1e257bbf8a-ovsdbserver-nb\") pod \"ae2f3329-33ca-402d-b841-8f1e257bbf8a\" (UID: \"ae2f3329-33ca-402d-b841-8f1e257bbf8a\") " Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.599499 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ae2f3329-33ca-402d-b841-8f1e257bbf8a-dns-swift-storage-0\") pod \"ae2f3329-33ca-402d-b841-8f1e257bbf8a\" (UID: \"ae2f3329-33ca-402d-b841-8f1e257bbf8a\") " Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.605726 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae2f3329-33ca-402d-b841-8f1e257bbf8a-kube-api-access-nrvf9" (OuterVolumeSpecName: "kube-api-access-nrvf9") pod "ae2f3329-33ca-402d-b841-8f1e257bbf8a" (UID: "ae2f3329-33ca-402d-b841-8f1e257bbf8a"). InnerVolumeSpecName "kube-api-access-nrvf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.650229 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae2f3329-33ca-402d-b841-8f1e257bbf8a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ae2f3329-33ca-402d-b841-8f1e257bbf8a" (UID: "ae2f3329-33ca-402d-b841-8f1e257bbf8a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.653772 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae2f3329-33ca-402d-b841-8f1e257bbf8a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ae2f3329-33ca-402d-b841-8f1e257bbf8a" (UID: "ae2f3329-33ca-402d-b841-8f1e257bbf8a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.684131 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae2f3329-33ca-402d-b841-8f1e257bbf8a-config" (OuterVolumeSpecName: "config") pod "ae2f3329-33ca-402d-b841-8f1e257bbf8a" (UID: "ae2f3329-33ca-402d-b841-8f1e257bbf8a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.686718 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae2f3329-33ca-402d-b841-8f1e257bbf8a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ae2f3329-33ca-402d-b841-8f1e257bbf8a" (UID: "ae2f3329-33ca-402d-b841-8f1e257bbf8a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.686905 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae2f3329-33ca-402d-b841-8f1e257bbf8a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ae2f3329-33ca-402d-b841-8f1e257bbf8a" (UID: "ae2f3329-33ca-402d-b841-8f1e257bbf8a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.702089 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrvf9\" (UniqueName: \"kubernetes.io/projected/ae2f3329-33ca-402d-b841-8f1e257bbf8a-kube-api-access-nrvf9\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.702500 4840 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae2f3329-33ca-402d-b841-8f1e257bbf8a-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.702517 4840 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae2f3329-33ca-402d-b841-8f1e257bbf8a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.702529 4840 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ae2f3329-33ca-402d-b841-8f1e257bbf8a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.702545 4840 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ae2f3329-33ca-402d-b841-8f1e257bbf8a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:21 crc kubenswrapper[4840]: I0930 14:15:21.702600 4840 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae2f3329-33ca-402d-b841-8f1e257bbf8a-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:22 crc kubenswrapper[4840]: I0930 14:15:22.043775 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 14:15:22 crc kubenswrapper[4840]: W0930 14:15:22.048462 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda14c8d91_8aff_4a17_8c9f_ce3839b3dbed.slice/crio-527bb14db4a24724d0ce9906203a5576bd5eea878813d187f574b204c7835ae5 WatchSource:0}: Error finding container 527bb14db4a24724d0ce9906203a5576bd5eea878813d187f574b204c7835ae5: Status 404 returned error can't find the container with id 527bb14db4a24724d0ce9906203a5576bd5eea878813d187f574b204c7835ae5 Sep 30 14:15:22 crc kubenswrapper[4840]: I0930 14:15:22.063318 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a14c8d91-8aff-4a17-8c9f-ce3839b3dbed","Type":"ContainerStarted","Data":"527bb14db4a24724d0ce9906203a5576bd5eea878813d187f574b204c7835ae5"} Sep 30 14:15:22 crc kubenswrapper[4840]: I0930 14:15:22.066440 4840 generic.go:334] "Generic (PLEG): container finished" podID="712feb06-cb97-425d-8b75-fcd5815a94ab" containerID="9ffdf9131bd8af23b5cc76199769f964619ded52d0516748629ec0e503506d3c" exitCode=0 Sep 30 14:15:22 crc kubenswrapper[4840]: I0930 14:15:22.066521 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"712feb06-cb97-425d-8b75-fcd5815a94ab","Type":"ContainerDied","Data":"9ffdf9131bd8af23b5cc76199769f964619ded52d0516748629ec0e503506d3c"} Sep 30 14:15:22 crc kubenswrapper[4840]: I0930 14:15:22.068423 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-qs9s6" event={"ID":"ae2f3329-33ca-402d-b841-8f1e257bbf8a","Type":"ContainerDied","Data":"6789707326d38b24e2609c00bdd94a43191af819d0e1ad323c450ef1f245470b"} Sep 30 14:15:22 crc kubenswrapper[4840]: I0930 14:15:22.068464 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-qs9s6" Sep 30 14:15:22 crc kubenswrapper[4840]: I0930 14:15:22.068495 4840 scope.go:117] "RemoveContainer" containerID="71b5753eee0c71364cb76a4843975d6ba07a09e3c6338c5ac66b32b4a82d65d0" Sep 30 14:15:22 crc kubenswrapper[4840]: I0930 14:15:22.068538 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="e616a00a-2b1e-4ea6-bde6-1a5517c8464e" containerName="nova-scheduler-scheduler" containerID="cri-o://d0a85ca9716a373dcf69f3c04d2251d1426d5a6fadaaf28100c017ceffd8affb" gracePeriod=30 Sep 30 14:15:22 crc kubenswrapper[4840]: I0930 14:15:22.106757 4840 scope.go:117] "RemoveContainer" containerID="218b0c923a2f1127faf075a7b6fae3b7c0e6693b90820678f8522b3ef476766a" Sep 30 14:15:22 crc kubenswrapper[4840]: I0930 14:15:22.107758 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-qs9s6"] Sep 30 14:15:22 crc kubenswrapper[4840]: I0930 14:15:22.131175 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70d820fe-33fc-4d18-b75f-3bff2044ba5b" path="/var/lib/kubelet/pods/70d820fe-33fc-4d18-b75f-3bff2044ba5b/volumes" Sep 30 14:15:22 crc kubenswrapper[4840]: I0930 14:15:22.131744 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-qs9s6"] Sep 30 14:15:23 crc kubenswrapper[4840]: E0930 14:15:23.404229 4840 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d0a85ca9716a373dcf69f3c04d2251d1426d5a6fadaaf28100c017ceffd8affb" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Sep 30 14:15:23 crc kubenswrapper[4840]: E0930 14:15:23.409805 4840 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d0a85ca9716a373dcf69f3c04d2251d1426d5a6fadaaf28100c017ceffd8affb" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Sep 30 14:15:23 crc kubenswrapper[4840]: E0930 14:15:23.411377 4840 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d0a85ca9716a373dcf69f3c04d2251d1426d5a6fadaaf28100c017ceffd8affb" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Sep 30 14:15:23 crc kubenswrapper[4840]: E0930 14:15:23.411416 4840 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="e616a00a-2b1e-4ea6-bde6-1a5517c8464e" containerName="nova-scheduler-scheduler" Sep 30 14:15:24 crc kubenswrapper[4840]: I0930 14:15:24.090179 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a14c8d91-8aff-4a17-8c9f-ce3839b3dbed","Type":"ContainerStarted","Data":"f460bec6a12f70ed6c569f9cac301dc1f74fe1dd0ef1a5be64f6c86939b7918b"} Sep 30 14:15:24 crc kubenswrapper[4840]: I0930 14:15:24.093093 4840 generic.go:334] "Generic (PLEG): container finished" podID="712feb06-cb97-425d-8b75-fcd5815a94ab" containerID="fa186597d4025b0b4a5b61ec12ce99d1733596a1825f242ee2bcbfc59b2000c6" exitCode=0 Sep 30 14:15:24 crc kubenswrapper[4840]: I0930 14:15:24.093125 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"712feb06-cb97-425d-8b75-fcd5815a94ab","Type":"ContainerDied","Data":"fa186597d4025b0b4a5b61ec12ce99d1733596a1825f242ee2bcbfc59b2000c6"} Sep 30 14:15:24 crc kubenswrapper[4840]: I0930 14:15:24.128171 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae2f3329-33ca-402d-b841-8f1e257bbf8a" path="/var/lib/kubelet/pods/ae2f3329-33ca-402d-b841-8f1e257bbf8a/volumes" Sep 30 14:15:24 crc kubenswrapper[4840]: I0930 14:15:24.529157 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:15:24 crc kubenswrapper[4840]: I0930 14:15:24.652088 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/712feb06-cb97-425d-8b75-fcd5815a94ab-run-httpd\") pod \"712feb06-cb97-425d-8b75-fcd5815a94ab\" (UID: \"712feb06-cb97-425d-8b75-fcd5815a94ab\") " Sep 30 14:15:24 crc kubenswrapper[4840]: I0930 14:15:24.652403 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/712feb06-cb97-425d-8b75-fcd5815a94ab-sg-core-conf-yaml\") pod \"712feb06-cb97-425d-8b75-fcd5815a94ab\" (UID: \"712feb06-cb97-425d-8b75-fcd5815a94ab\") " Sep 30 14:15:24 crc kubenswrapper[4840]: I0930 14:15:24.652463 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/712feb06-cb97-425d-8b75-fcd5815a94ab-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "712feb06-cb97-425d-8b75-fcd5815a94ab" (UID: "712feb06-cb97-425d-8b75-fcd5815a94ab"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:15:24 crc kubenswrapper[4840]: I0930 14:15:24.652499 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rsh6k\" (UniqueName: \"kubernetes.io/projected/712feb06-cb97-425d-8b75-fcd5815a94ab-kube-api-access-rsh6k\") pod \"712feb06-cb97-425d-8b75-fcd5815a94ab\" (UID: \"712feb06-cb97-425d-8b75-fcd5815a94ab\") " Sep 30 14:15:24 crc kubenswrapper[4840]: I0930 14:15:24.652576 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/712feb06-cb97-425d-8b75-fcd5815a94ab-scripts\") pod \"712feb06-cb97-425d-8b75-fcd5815a94ab\" (UID: \"712feb06-cb97-425d-8b75-fcd5815a94ab\") " Sep 30 14:15:24 crc kubenswrapper[4840]: I0930 14:15:24.652606 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/712feb06-cb97-425d-8b75-fcd5815a94ab-combined-ca-bundle\") pod \"712feb06-cb97-425d-8b75-fcd5815a94ab\" (UID: \"712feb06-cb97-425d-8b75-fcd5815a94ab\") " Sep 30 14:15:24 crc kubenswrapper[4840]: I0930 14:15:24.652658 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/712feb06-cb97-425d-8b75-fcd5815a94ab-config-data\") pod \"712feb06-cb97-425d-8b75-fcd5815a94ab\" (UID: \"712feb06-cb97-425d-8b75-fcd5815a94ab\") " Sep 30 14:15:24 crc kubenswrapper[4840]: I0930 14:15:24.652742 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/712feb06-cb97-425d-8b75-fcd5815a94ab-log-httpd\") pod \"712feb06-cb97-425d-8b75-fcd5815a94ab\" (UID: \"712feb06-cb97-425d-8b75-fcd5815a94ab\") " Sep 30 14:15:24 crc kubenswrapper[4840]: I0930 14:15:24.653090 4840 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/712feb06-cb97-425d-8b75-fcd5815a94ab-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:24 crc kubenswrapper[4840]: I0930 14:15:24.653450 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/712feb06-cb97-425d-8b75-fcd5815a94ab-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "712feb06-cb97-425d-8b75-fcd5815a94ab" (UID: "712feb06-cb97-425d-8b75-fcd5815a94ab"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:15:24 crc kubenswrapper[4840]: I0930 14:15:24.657528 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/712feb06-cb97-425d-8b75-fcd5815a94ab-scripts" (OuterVolumeSpecName: "scripts") pod "712feb06-cb97-425d-8b75-fcd5815a94ab" (UID: "712feb06-cb97-425d-8b75-fcd5815a94ab"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:15:24 crc kubenswrapper[4840]: I0930 14:15:24.658264 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/712feb06-cb97-425d-8b75-fcd5815a94ab-kube-api-access-rsh6k" (OuterVolumeSpecName: "kube-api-access-rsh6k") pod "712feb06-cb97-425d-8b75-fcd5815a94ab" (UID: "712feb06-cb97-425d-8b75-fcd5815a94ab"). InnerVolumeSpecName "kube-api-access-rsh6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:15:24 crc kubenswrapper[4840]: I0930 14:15:24.682994 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/712feb06-cb97-425d-8b75-fcd5815a94ab-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "712feb06-cb97-425d-8b75-fcd5815a94ab" (UID: "712feb06-cb97-425d-8b75-fcd5815a94ab"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:15:24 crc kubenswrapper[4840]: I0930 14:15:24.746289 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/712feb06-cb97-425d-8b75-fcd5815a94ab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "712feb06-cb97-425d-8b75-fcd5815a94ab" (UID: "712feb06-cb97-425d-8b75-fcd5815a94ab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:15:24 crc kubenswrapper[4840]: I0930 14:15:24.754919 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rsh6k\" (UniqueName: \"kubernetes.io/projected/712feb06-cb97-425d-8b75-fcd5815a94ab-kube-api-access-rsh6k\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:24 crc kubenswrapper[4840]: I0930 14:15:24.755503 4840 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/712feb06-cb97-425d-8b75-fcd5815a94ab-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:24 crc kubenswrapper[4840]: I0930 14:15:24.755957 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/712feb06-cb97-425d-8b75-fcd5815a94ab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:24 crc kubenswrapper[4840]: I0930 14:15:24.756054 4840 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/712feb06-cb97-425d-8b75-fcd5815a94ab-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:24 crc kubenswrapper[4840]: I0930 14:15:24.756126 4840 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/712feb06-cb97-425d-8b75-fcd5815a94ab-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:24 crc kubenswrapper[4840]: I0930 14:15:24.762424 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/712feb06-cb97-425d-8b75-fcd5815a94ab-config-data" (OuterVolumeSpecName: "config-data") pod "712feb06-cb97-425d-8b75-fcd5815a94ab" (UID: "712feb06-cb97-425d-8b75-fcd5815a94ab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:15:24 crc kubenswrapper[4840]: I0930 14:15:24.858018 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/712feb06-cb97-425d-8b75-fcd5815a94ab-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.112032 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a14c8d91-8aff-4a17-8c9f-ce3839b3dbed","Type":"ContainerStarted","Data":"acece978aba77ed13654be82a29a690286db631f6fa052a2922f109a7b51e675"} Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.116440 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.116474 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"712feb06-cb97-425d-8b75-fcd5815a94ab","Type":"ContainerDied","Data":"a0ed70585198d5191ed48940c56b05d8a544bb56844542b03b3b43d683eb138f"} Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.116579 4840 scope.go:117] "RemoveContainer" containerID="25e57148f578d9ab12ef79020c7ac3c16f6b8542b50b2833a6d175c2d8a4c3f0" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.130123 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"26d4f9e8-9d18-45ec-8117-39c0714ce1b0","Type":"ContainerStarted","Data":"16286168d801e737b3e3c8bf612a256ba253770fb9234523c4b82718967bab58"} Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.130741 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.132530 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=4.1325127 podStartE2EDuration="4.1325127s" podCreationTimestamp="2025-09-30 14:15:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:15:25.129826 +0000 UTC m=+1153.758912443" watchObservedRunningTime="2025-09-30 14:15:25.1325127 +0000 UTC m=+1153.761599123" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.155305 4840 scope.go:117] "RemoveContainer" containerID="53aa98fbc16bc77ba4f16d0623177aa092f9f3f0a29df3651490494fd3510c5d" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.167344 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.611551805 podStartE2EDuration="6.167326867s" podCreationTimestamp="2025-09-30 14:15:19 +0000 UTC" firstStartedPulling="2025-09-30 14:15:19.905520203 +0000 UTC m=+1148.534606616" lastFinishedPulling="2025-09-30 14:15:24.461295255 +0000 UTC m=+1153.090381678" observedRunningTime="2025-09-30 14:15:25.160311494 +0000 UTC m=+1153.789397917" watchObservedRunningTime="2025-09-30 14:15:25.167326867 +0000 UTC m=+1153.796413290" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.180739 4840 scope.go:117] "RemoveContainer" containerID="fa186597d4025b0b4a5b61ec12ce99d1733596a1825f242ee2bcbfc59b2000c6" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.186635 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.193883 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.213295 4840 scope.go:117] "RemoveContainer" containerID="9ffdf9131bd8af23b5cc76199769f964619ded52d0516748629ec0e503506d3c" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.216171 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:15:25 crc kubenswrapper[4840]: E0930 14:15:25.217775 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae2f3329-33ca-402d-b841-8f1e257bbf8a" containerName="dnsmasq-dns" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.217801 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae2f3329-33ca-402d-b841-8f1e257bbf8a" containerName="dnsmasq-dns" Sep 30 14:15:25 crc kubenswrapper[4840]: E0930 14:15:25.217877 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae2f3329-33ca-402d-b841-8f1e257bbf8a" containerName="init" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.217887 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae2f3329-33ca-402d-b841-8f1e257bbf8a" containerName="init" Sep 30 14:15:25 crc kubenswrapper[4840]: E0930 14:15:25.217898 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="712feb06-cb97-425d-8b75-fcd5815a94ab" containerName="ceilometer-central-agent" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.217905 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="712feb06-cb97-425d-8b75-fcd5815a94ab" containerName="ceilometer-central-agent" Sep 30 14:15:25 crc kubenswrapper[4840]: E0930 14:15:25.217921 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="712feb06-cb97-425d-8b75-fcd5815a94ab" containerName="ceilometer-notification-agent" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.217928 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="712feb06-cb97-425d-8b75-fcd5815a94ab" containerName="ceilometer-notification-agent" Sep 30 14:15:25 crc kubenswrapper[4840]: E0930 14:15:25.217939 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="712feb06-cb97-425d-8b75-fcd5815a94ab" containerName="proxy-httpd" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.217947 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="712feb06-cb97-425d-8b75-fcd5815a94ab" containerName="proxy-httpd" Sep 30 14:15:25 crc kubenswrapper[4840]: E0930 14:15:25.218481 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="712feb06-cb97-425d-8b75-fcd5815a94ab" containerName="sg-core" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.218500 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="712feb06-cb97-425d-8b75-fcd5815a94ab" containerName="sg-core" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.218915 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="712feb06-cb97-425d-8b75-fcd5815a94ab" containerName="proxy-httpd" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.218945 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="712feb06-cb97-425d-8b75-fcd5815a94ab" containerName="sg-core" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.218985 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="712feb06-cb97-425d-8b75-fcd5815a94ab" containerName="ceilometer-notification-agent" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.219001 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="712feb06-cb97-425d-8b75-fcd5815a94ab" containerName="ceilometer-central-agent" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.219013 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae2f3329-33ca-402d-b841-8f1e257bbf8a" containerName="dnsmasq-dns" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.220922 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.224538 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.226924 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.227124 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.227307 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.368732 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4xsz\" (UniqueName: \"kubernetes.io/projected/96e1874a-e455-497d-8593-fd9f09bd39e3-kube-api-access-d4xsz\") pod \"ceilometer-0\" (UID: \"96e1874a-e455-497d-8593-fd9f09bd39e3\") " pod="openstack/ceilometer-0" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.369206 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96e1874a-e455-497d-8593-fd9f09bd39e3-scripts\") pod \"ceilometer-0\" (UID: \"96e1874a-e455-497d-8593-fd9f09bd39e3\") " pod="openstack/ceilometer-0" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.369260 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/96e1874a-e455-497d-8593-fd9f09bd39e3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"96e1874a-e455-497d-8593-fd9f09bd39e3\") " pod="openstack/ceilometer-0" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.369360 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96e1874a-e455-497d-8593-fd9f09bd39e3-config-data\") pod \"ceilometer-0\" (UID: \"96e1874a-e455-497d-8593-fd9f09bd39e3\") " pod="openstack/ceilometer-0" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.369405 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96e1874a-e455-497d-8593-fd9f09bd39e3-log-httpd\") pod \"ceilometer-0\" (UID: \"96e1874a-e455-497d-8593-fd9f09bd39e3\") " pod="openstack/ceilometer-0" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.369527 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96e1874a-e455-497d-8593-fd9f09bd39e3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"96e1874a-e455-497d-8593-fd9f09bd39e3\") " pod="openstack/ceilometer-0" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.369656 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96e1874a-e455-497d-8593-fd9f09bd39e3-run-httpd\") pod \"ceilometer-0\" (UID: \"96e1874a-e455-497d-8593-fd9f09bd39e3\") " pod="openstack/ceilometer-0" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.369706 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/96e1874a-e455-497d-8593-fd9f09bd39e3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"96e1874a-e455-497d-8593-fd9f09bd39e3\") " pod="openstack/ceilometer-0" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.471031 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96e1874a-e455-497d-8593-fd9f09bd39e3-run-httpd\") pod \"ceilometer-0\" (UID: \"96e1874a-e455-497d-8593-fd9f09bd39e3\") " pod="openstack/ceilometer-0" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.471077 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/96e1874a-e455-497d-8593-fd9f09bd39e3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"96e1874a-e455-497d-8593-fd9f09bd39e3\") " pod="openstack/ceilometer-0" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.471125 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4xsz\" (UniqueName: \"kubernetes.io/projected/96e1874a-e455-497d-8593-fd9f09bd39e3-kube-api-access-d4xsz\") pod \"ceilometer-0\" (UID: \"96e1874a-e455-497d-8593-fd9f09bd39e3\") " pod="openstack/ceilometer-0" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.471195 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96e1874a-e455-497d-8593-fd9f09bd39e3-scripts\") pod \"ceilometer-0\" (UID: \"96e1874a-e455-497d-8593-fd9f09bd39e3\") " pod="openstack/ceilometer-0" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.471230 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/96e1874a-e455-497d-8593-fd9f09bd39e3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"96e1874a-e455-497d-8593-fd9f09bd39e3\") " pod="openstack/ceilometer-0" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.471247 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96e1874a-e455-497d-8593-fd9f09bd39e3-config-data\") pod \"ceilometer-0\" (UID: \"96e1874a-e455-497d-8593-fd9f09bd39e3\") " pod="openstack/ceilometer-0" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.471259 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96e1874a-e455-497d-8593-fd9f09bd39e3-log-httpd\") pod \"ceilometer-0\" (UID: \"96e1874a-e455-497d-8593-fd9f09bd39e3\") " pod="openstack/ceilometer-0" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.471294 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96e1874a-e455-497d-8593-fd9f09bd39e3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"96e1874a-e455-497d-8593-fd9f09bd39e3\") " pod="openstack/ceilometer-0" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.472149 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96e1874a-e455-497d-8593-fd9f09bd39e3-run-httpd\") pod \"ceilometer-0\" (UID: \"96e1874a-e455-497d-8593-fd9f09bd39e3\") " pod="openstack/ceilometer-0" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.472369 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96e1874a-e455-497d-8593-fd9f09bd39e3-log-httpd\") pod \"ceilometer-0\" (UID: \"96e1874a-e455-497d-8593-fd9f09bd39e3\") " pod="openstack/ceilometer-0" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.475523 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96e1874a-e455-497d-8593-fd9f09bd39e3-config-data\") pod \"ceilometer-0\" (UID: \"96e1874a-e455-497d-8593-fd9f09bd39e3\") " pod="openstack/ceilometer-0" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.475522 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/96e1874a-e455-497d-8593-fd9f09bd39e3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"96e1874a-e455-497d-8593-fd9f09bd39e3\") " pod="openstack/ceilometer-0" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.475808 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/96e1874a-e455-497d-8593-fd9f09bd39e3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"96e1874a-e455-497d-8593-fd9f09bd39e3\") " pod="openstack/ceilometer-0" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.475981 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96e1874a-e455-497d-8593-fd9f09bd39e3-scripts\") pod \"ceilometer-0\" (UID: \"96e1874a-e455-497d-8593-fd9f09bd39e3\") " pod="openstack/ceilometer-0" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.477746 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96e1874a-e455-497d-8593-fd9f09bd39e3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"96e1874a-e455-497d-8593-fd9f09bd39e3\") " pod="openstack/ceilometer-0" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.489726 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4xsz\" (UniqueName: \"kubernetes.io/projected/96e1874a-e455-497d-8593-fd9f09bd39e3-kube-api-access-d4xsz\") pod \"ceilometer-0\" (UID: \"96e1874a-e455-497d-8593-fd9f09bd39e3\") " pod="openstack/ceilometer-0" Sep 30 14:15:25 crc kubenswrapper[4840]: I0930 14:15:25.543139 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.048222 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.137891 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="712feb06-cb97-425d-8b75-fcd5815a94ab" path="/var/lib/kubelet/pods/712feb06-cb97-425d-8b75-fcd5815a94ab/volumes" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.148181 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.157020 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96e1874a-e455-497d-8593-fd9f09bd39e3","Type":"ContainerStarted","Data":"cbb2d50ac7a03cdf6b68437c33714eb624e780cce6bd6e1b9e0496013ad3a9d6"} Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.158688 4840 generic.go:334] "Generic (PLEG): container finished" podID="e616a00a-2b1e-4ea6-bde6-1a5517c8464e" containerID="d0a85ca9716a373dcf69f3c04d2251d1426d5a6fadaaf28100c017ceffd8affb" exitCode=0 Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.158750 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e616a00a-2b1e-4ea6-bde6-1a5517c8464e","Type":"ContainerDied","Data":"d0a85ca9716a373dcf69f3c04d2251d1426d5a6fadaaf28100c017ceffd8affb"} Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.162006 4840 generic.go:334] "Generic (PLEG): container finished" podID="bc04dbcf-177f-4039-be37-9986232e2d42" containerID="d0e85f6c23c257ef6bbd38b05fea7e9d3832058240ab6ae45ee6c8d982f17c47" exitCode=0 Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.162270 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"bc04dbcf-177f-4039-be37-9986232e2d42","Type":"ContainerDied","Data":"d0e85f6c23c257ef6bbd38b05fea7e9d3832058240ab6ae45ee6c8d982f17c47"} Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.162297 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"bc04dbcf-177f-4039-be37-9986232e2d42","Type":"ContainerDied","Data":"7048570c998f9562d67033e22a65838b625dd1d6034b1449b6ce82318d28de6a"} Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.162312 4840 scope.go:117] "RemoveContainer" containerID="d0e85f6c23c257ef6bbd38b05fea7e9d3832058240ab6ae45ee6c8d982f17c47" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.162693 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.185430 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5784cf869f-qs9s6" podUID="ae2f3329-33ca-402d-b841-8f1e257bbf8a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.164:5353: i/o timeout" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.203683 4840 scope.go:117] "RemoveContainer" containerID="2c0a5ef199487715a5f0c8297b85a8e37514dd7d080cbaa9715e7083c1447fa7" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.267020 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.277316 4840 scope.go:117] "RemoveContainer" containerID="d0e85f6c23c257ef6bbd38b05fea7e9d3832058240ab6ae45ee6c8d982f17c47" Sep 30 14:15:26 crc kubenswrapper[4840]: E0930 14:15:26.277937 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0e85f6c23c257ef6bbd38b05fea7e9d3832058240ab6ae45ee6c8d982f17c47\": container with ID starting with d0e85f6c23c257ef6bbd38b05fea7e9d3832058240ab6ae45ee6c8d982f17c47 not found: ID does not exist" containerID="d0e85f6c23c257ef6bbd38b05fea7e9d3832058240ab6ae45ee6c8d982f17c47" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.278087 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0e85f6c23c257ef6bbd38b05fea7e9d3832058240ab6ae45ee6c8d982f17c47"} err="failed to get container status \"d0e85f6c23c257ef6bbd38b05fea7e9d3832058240ab6ae45ee6c8d982f17c47\": rpc error: code = NotFound desc = could not find container \"d0e85f6c23c257ef6bbd38b05fea7e9d3832058240ab6ae45ee6c8d982f17c47\": container with ID starting with d0e85f6c23c257ef6bbd38b05fea7e9d3832058240ab6ae45ee6c8d982f17c47 not found: ID does not exist" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.278204 4840 scope.go:117] "RemoveContainer" containerID="2c0a5ef199487715a5f0c8297b85a8e37514dd7d080cbaa9715e7083c1447fa7" Sep 30 14:15:26 crc kubenswrapper[4840]: E0930 14:15:26.278715 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c0a5ef199487715a5f0c8297b85a8e37514dd7d080cbaa9715e7083c1447fa7\": container with ID starting with 2c0a5ef199487715a5f0c8297b85a8e37514dd7d080cbaa9715e7083c1447fa7 not found: ID does not exist" containerID="2c0a5ef199487715a5f0c8297b85a8e37514dd7d080cbaa9715e7083c1447fa7" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.278735 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c0a5ef199487715a5f0c8297b85a8e37514dd7d080cbaa9715e7083c1447fa7"} err="failed to get container status \"2c0a5ef199487715a5f0c8297b85a8e37514dd7d080cbaa9715e7083c1447fa7\": rpc error: code = NotFound desc = could not find container \"2c0a5ef199487715a5f0c8297b85a8e37514dd7d080cbaa9715e7083c1447fa7\": container with ID starting with 2c0a5ef199487715a5f0c8297b85a8e37514dd7d080cbaa9715e7083c1447fa7 not found: ID does not exist" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.286749 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc04dbcf-177f-4039-be37-9986232e2d42-config-data\") pod \"bc04dbcf-177f-4039-be37-9986232e2d42\" (UID: \"bc04dbcf-177f-4039-be37-9986232e2d42\") " Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.286976 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc04dbcf-177f-4039-be37-9986232e2d42-logs\") pod \"bc04dbcf-177f-4039-be37-9986232e2d42\" (UID: \"bc04dbcf-177f-4039-be37-9986232e2d42\") " Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.287024 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc04dbcf-177f-4039-be37-9986232e2d42-combined-ca-bundle\") pod \"bc04dbcf-177f-4039-be37-9986232e2d42\" (UID: \"bc04dbcf-177f-4039-be37-9986232e2d42\") " Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.287083 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2kz5\" (UniqueName: \"kubernetes.io/projected/bc04dbcf-177f-4039-be37-9986232e2d42-kube-api-access-h2kz5\") pod \"bc04dbcf-177f-4039-be37-9986232e2d42\" (UID: \"bc04dbcf-177f-4039-be37-9986232e2d42\") " Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.287442 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc04dbcf-177f-4039-be37-9986232e2d42-logs" (OuterVolumeSpecName: "logs") pod "bc04dbcf-177f-4039-be37-9986232e2d42" (UID: "bc04dbcf-177f-4039-be37-9986232e2d42"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.294466 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc04dbcf-177f-4039-be37-9986232e2d42-kube-api-access-h2kz5" (OuterVolumeSpecName: "kube-api-access-h2kz5") pod "bc04dbcf-177f-4039-be37-9986232e2d42" (UID: "bc04dbcf-177f-4039-be37-9986232e2d42"). InnerVolumeSpecName "kube-api-access-h2kz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.328796 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc04dbcf-177f-4039-be37-9986232e2d42-config-data" (OuterVolumeSpecName: "config-data") pod "bc04dbcf-177f-4039-be37-9986232e2d42" (UID: "bc04dbcf-177f-4039-be37-9986232e2d42"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.338146 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc04dbcf-177f-4039-be37-9986232e2d42-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bc04dbcf-177f-4039-be37-9986232e2d42" (UID: "bc04dbcf-177f-4039-be37-9986232e2d42"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.388797 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e616a00a-2b1e-4ea6-bde6-1a5517c8464e-combined-ca-bundle\") pod \"e616a00a-2b1e-4ea6-bde6-1a5517c8464e\" (UID: \"e616a00a-2b1e-4ea6-bde6-1a5517c8464e\") " Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.389003 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e616a00a-2b1e-4ea6-bde6-1a5517c8464e-config-data\") pod \"e616a00a-2b1e-4ea6-bde6-1a5517c8464e\" (UID: \"e616a00a-2b1e-4ea6-bde6-1a5517c8464e\") " Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.389071 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjxwk\" (UniqueName: \"kubernetes.io/projected/e616a00a-2b1e-4ea6-bde6-1a5517c8464e-kube-api-access-sjxwk\") pod \"e616a00a-2b1e-4ea6-bde6-1a5517c8464e\" (UID: \"e616a00a-2b1e-4ea6-bde6-1a5517c8464e\") " Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.389852 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc04dbcf-177f-4039-be37-9986232e2d42-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.389875 4840 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc04dbcf-177f-4039-be37-9986232e2d42-logs\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.389885 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc04dbcf-177f-4039-be37-9986232e2d42-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.389896 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2kz5\" (UniqueName: \"kubernetes.io/projected/bc04dbcf-177f-4039-be37-9986232e2d42-kube-api-access-h2kz5\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.392781 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e616a00a-2b1e-4ea6-bde6-1a5517c8464e-kube-api-access-sjxwk" (OuterVolumeSpecName: "kube-api-access-sjxwk") pod "e616a00a-2b1e-4ea6-bde6-1a5517c8464e" (UID: "e616a00a-2b1e-4ea6-bde6-1a5517c8464e"). InnerVolumeSpecName "kube-api-access-sjxwk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.427775 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e616a00a-2b1e-4ea6-bde6-1a5517c8464e-config-data" (OuterVolumeSpecName: "config-data") pod "e616a00a-2b1e-4ea6-bde6-1a5517c8464e" (UID: "e616a00a-2b1e-4ea6-bde6-1a5517c8464e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.430263 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e616a00a-2b1e-4ea6-bde6-1a5517c8464e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e616a00a-2b1e-4ea6-bde6-1a5517c8464e" (UID: "e616a00a-2b1e-4ea6-bde6-1a5517c8464e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.477709 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.477769 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.492460 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e616a00a-2b1e-4ea6-bde6-1a5517c8464e-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.492491 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjxwk\" (UniqueName: \"kubernetes.io/projected/e616a00a-2b1e-4ea6-bde6-1a5517c8464e-kube-api-access-sjxwk\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.492502 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e616a00a-2b1e-4ea6-bde6-1a5517c8464e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.497727 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.505208 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.515723 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Sep 30 14:15:26 crc kubenswrapper[4840]: E0930 14:15:26.516090 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc04dbcf-177f-4039-be37-9986232e2d42" containerName="nova-api-api" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.516110 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc04dbcf-177f-4039-be37-9986232e2d42" containerName="nova-api-api" Sep 30 14:15:26 crc kubenswrapper[4840]: E0930 14:15:26.516122 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e616a00a-2b1e-4ea6-bde6-1a5517c8464e" containerName="nova-scheduler-scheduler" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.516129 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="e616a00a-2b1e-4ea6-bde6-1a5517c8464e" containerName="nova-scheduler-scheduler" Sep 30 14:15:26 crc kubenswrapper[4840]: E0930 14:15:26.516171 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc04dbcf-177f-4039-be37-9986232e2d42" containerName="nova-api-log" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.516177 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc04dbcf-177f-4039-be37-9986232e2d42" containerName="nova-api-log" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.516328 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc04dbcf-177f-4039-be37-9986232e2d42" containerName="nova-api-api" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.516338 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="e616a00a-2b1e-4ea6-bde6-1a5517c8464e" containerName="nova-scheduler-scheduler" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.516352 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc04dbcf-177f-4039-be37-9986232e2d42" containerName="nova-api-log" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.517770 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.523159 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.529087 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.593875 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpxxf\" (UniqueName: \"kubernetes.io/projected/2583ecd5-01f5-4e7d-907b-9dc8a3771273-kube-api-access-tpxxf\") pod \"nova-api-0\" (UID: \"2583ecd5-01f5-4e7d-907b-9dc8a3771273\") " pod="openstack/nova-api-0" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.594164 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2583ecd5-01f5-4e7d-907b-9dc8a3771273-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2583ecd5-01f5-4e7d-907b-9dc8a3771273\") " pod="openstack/nova-api-0" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.594221 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2583ecd5-01f5-4e7d-907b-9dc8a3771273-logs\") pod \"nova-api-0\" (UID: \"2583ecd5-01f5-4e7d-907b-9dc8a3771273\") " pod="openstack/nova-api-0" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.594266 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2583ecd5-01f5-4e7d-907b-9dc8a3771273-config-data\") pod \"nova-api-0\" (UID: \"2583ecd5-01f5-4e7d-907b-9dc8a3771273\") " pod="openstack/nova-api-0" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.696178 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2583ecd5-01f5-4e7d-907b-9dc8a3771273-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2583ecd5-01f5-4e7d-907b-9dc8a3771273\") " pod="openstack/nova-api-0" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.696237 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2583ecd5-01f5-4e7d-907b-9dc8a3771273-logs\") pod \"nova-api-0\" (UID: \"2583ecd5-01f5-4e7d-907b-9dc8a3771273\") " pod="openstack/nova-api-0" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.696267 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2583ecd5-01f5-4e7d-907b-9dc8a3771273-config-data\") pod \"nova-api-0\" (UID: \"2583ecd5-01f5-4e7d-907b-9dc8a3771273\") " pod="openstack/nova-api-0" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.696318 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpxxf\" (UniqueName: \"kubernetes.io/projected/2583ecd5-01f5-4e7d-907b-9dc8a3771273-kube-api-access-tpxxf\") pod \"nova-api-0\" (UID: \"2583ecd5-01f5-4e7d-907b-9dc8a3771273\") " pod="openstack/nova-api-0" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.697564 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2583ecd5-01f5-4e7d-907b-9dc8a3771273-logs\") pod \"nova-api-0\" (UID: \"2583ecd5-01f5-4e7d-907b-9dc8a3771273\") " pod="openstack/nova-api-0" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.700183 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2583ecd5-01f5-4e7d-907b-9dc8a3771273-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2583ecd5-01f5-4e7d-907b-9dc8a3771273\") " pod="openstack/nova-api-0" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.700800 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2583ecd5-01f5-4e7d-907b-9dc8a3771273-config-data\") pod \"nova-api-0\" (UID: \"2583ecd5-01f5-4e7d-907b-9dc8a3771273\") " pod="openstack/nova-api-0" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.713994 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpxxf\" (UniqueName: \"kubernetes.io/projected/2583ecd5-01f5-4e7d-907b-9dc8a3771273-kube-api-access-tpxxf\") pod \"nova-api-0\" (UID: \"2583ecd5-01f5-4e7d-907b-9dc8a3771273\") " pod="openstack/nova-api-0" Sep 30 14:15:26 crc kubenswrapper[4840]: I0930 14:15:26.844071 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 14:15:27 crc kubenswrapper[4840]: I0930 14:15:27.177255 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 14:15:27 crc kubenswrapper[4840]: I0930 14:15:27.177285 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e616a00a-2b1e-4ea6-bde6-1a5517c8464e","Type":"ContainerDied","Data":"2999182eef0c988cb9dec38883198a1f9d47f049d14417495090236013dfef1a"} Sep 30 14:15:27 crc kubenswrapper[4840]: I0930 14:15:27.178614 4840 scope.go:117] "RemoveContainer" containerID="d0a85ca9716a373dcf69f3c04d2251d1426d5a6fadaaf28100c017ceffd8affb" Sep 30 14:15:27 crc kubenswrapper[4840]: I0930 14:15:27.213797 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 14:15:27 crc kubenswrapper[4840]: I0930 14:15:27.232398 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 14:15:27 crc kubenswrapper[4840]: I0930 14:15:27.244116 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 14:15:27 crc kubenswrapper[4840]: I0930 14:15:27.245405 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 14:15:27 crc kubenswrapper[4840]: I0930 14:15:27.247812 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Sep 30 14:15:27 crc kubenswrapper[4840]: I0930 14:15:27.252953 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 14:15:27 crc kubenswrapper[4840]: I0930 14:15:27.298237 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 30 14:15:27 crc kubenswrapper[4840]: W0930 14:15:27.299001 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2583ecd5_01f5_4e7d_907b_9dc8a3771273.slice/crio-68c87448b25d4f1cdbc875001fc5684a703a033ab1d0778c69a932f4f3ca64a1 WatchSource:0}: Error finding container 68c87448b25d4f1cdbc875001fc5684a703a033ab1d0778c69a932f4f3ca64a1: Status 404 returned error can't find the container with id 68c87448b25d4f1cdbc875001fc5684a703a033ab1d0778c69a932f4f3ca64a1 Sep 30 14:15:27 crc kubenswrapper[4840]: I0930 14:15:27.314762 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1658ae4f-b344-4f58-9b37-ba03da92230f-config-data\") pod \"nova-scheduler-0\" (UID: \"1658ae4f-b344-4f58-9b37-ba03da92230f\") " pod="openstack/nova-scheduler-0" Sep 30 14:15:27 crc kubenswrapper[4840]: I0930 14:15:27.314889 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1658ae4f-b344-4f58-9b37-ba03da92230f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1658ae4f-b344-4f58-9b37-ba03da92230f\") " pod="openstack/nova-scheduler-0" Sep 30 14:15:27 crc kubenswrapper[4840]: I0930 14:15:27.314996 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2b494\" (UniqueName: \"kubernetes.io/projected/1658ae4f-b344-4f58-9b37-ba03da92230f-kube-api-access-2b494\") pod \"nova-scheduler-0\" (UID: \"1658ae4f-b344-4f58-9b37-ba03da92230f\") " pod="openstack/nova-scheduler-0" Sep 30 14:15:27 crc kubenswrapper[4840]: I0930 14:15:27.416966 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2b494\" (UniqueName: \"kubernetes.io/projected/1658ae4f-b344-4f58-9b37-ba03da92230f-kube-api-access-2b494\") pod \"nova-scheduler-0\" (UID: \"1658ae4f-b344-4f58-9b37-ba03da92230f\") " pod="openstack/nova-scheduler-0" Sep 30 14:15:27 crc kubenswrapper[4840]: I0930 14:15:27.417370 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1658ae4f-b344-4f58-9b37-ba03da92230f-config-data\") pod \"nova-scheduler-0\" (UID: \"1658ae4f-b344-4f58-9b37-ba03da92230f\") " pod="openstack/nova-scheduler-0" Sep 30 14:15:27 crc kubenswrapper[4840]: I0930 14:15:27.417431 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1658ae4f-b344-4f58-9b37-ba03da92230f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1658ae4f-b344-4f58-9b37-ba03da92230f\") " pod="openstack/nova-scheduler-0" Sep 30 14:15:27 crc kubenswrapper[4840]: I0930 14:15:27.422610 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1658ae4f-b344-4f58-9b37-ba03da92230f-config-data\") pod \"nova-scheduler-0\" (UID: \"1658ae4f-b344-4f58-9b37-ba03da92230f\") " pod="openstack/nova-scheduler-0" Sep 30 14:15:27 crc kubenswrapper[4840]: I0930 14:15:27.422787 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1658ae4f-b344-4f58-9b37-ba03da92230f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1658ae4f-b344-4f58-9b37-ba03da92230f\") " pod="openstack/nova-scheduler-0" Sep 30 14:15:27 crc kubenswrapper[4840]: I0930 14:15:27.435015 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2b494\" (UniqueName: \"kubernetes.io/projected/1658ae4f-b344-4f58-9b37-ba03da92230f-kube-api-access-2b494\") pod \"nova-scheduler-0\" (UID: \"1658ae4f-b344-4f58-9b37-ba03da92230f\") " pod="openstack/nova-scheduler-0" Sep 30 14:15:27 crc kubenswrapper[4840]: I0930 14:15:27.581236 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 14:15:28 crc kubenswrapper[4840]: I0930 14:15:28.023208 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 14:15:28 crc kubenswrapper[4840]: I0930 14:15:28.135372 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc04dbcf-177f-4039-be37-9986232e2d42" path="/var/lib/kubelet/pods/bc04dbcf-177f-4039-be37-9986232e2d42/volumes" Sep 30 14:15:28 crc kubenswrapper[4840]: I0930 14:15:28.136625 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e616a00a-2b1e-4ea6-bde6-1a5517c8464e" path="/var/lib/kubelet/pods/e616a00a-2b1e-4ea6-bde6-1a5517c8464e/volumes" Sep 30 14:15:28 crc kubenswrapper[4840]: I0930 14:15:28.198096 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96e1874a-e455-497d-8593-fd9f09bd39e3","Type":"ContainerStarted","Data":"e748188bb7802680970886e08f18bf5277da7e67c46d1ac73a10235551660bb5"} Sep 30 14:15:28 crc kubenswrapper[4840]: I0930 14:15:28.200215 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1658ae4f-b344-4f58-9b37-ba03da92230f","Type":"ContainerStarted","Data":"d2ea6fa32ccbb46846d1c3682f161d6544c4a5b55c470b58269eac5301d3f6a8"} Sep 30 14:15:28 crc kubenswrapper[4840]: I0930 14:15:28.204095 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2583ecd5-01f5-4e7d-907b-9dc8a3771273","Type":"ContainerStarted","Data":"7ab0e5f8e7af49a56cc18f763c4df49e9d070427438a08aff82d02d580adb13b"} Sep 30 14:15:28 crc kubenswrapper[4840]: I0930 14:15:28.204135 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2583ecd5-01f5-4e7d-907b-9dc8a3771273","Type":"ContainerStarted","Data":"4afc1cadc812e77af8b425f63eb80d4af24749da13a5d82629a2e1d512110e57"} Sep 30 14:15:28 crc kubenswrapper[4840]: I0930 14:15:28.204145 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2583ecd5-01f5-4e7d-907b-9dc8a3771273","Type":"ContainerStarted","Data":"68c87448b25d4f1cdbc875001fc5684a703a033ab1d0778c69a932f4f3ca64a1"} Sep 30 14:15:29 crc kubenswrapper[4840]: I0930 14:15:29.217179 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96e1874a-e455-497d-8593-fd9f09bd39e3","Type":"ContainerStarted","Data":"27a04dcb560f3443c2b1c2bf943cac491fbc6a58af74155372c5516866e459ab"} Sep 30 14:15:29 crc kubenswrapper[4840]: I0930 14:15:29.219647 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1658ae4f-b344-4f58-9b37-ba03da92230f","Type":"ContainerStarted","Data":"4b5bd4ba815125f172e4e48d936f0189b8e3ee03ecdc094f46ecfad486aaa61b"} Sep 30 14:15:29 crc kubenswrapper[4840]: I0930 14:15:29.241486 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.241464064 podStartE2EDuration="2.241464064s" podCreationTimestamp="2025-09-30 14:15:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:15:29.238879847 +0000 UTC m=+1157.867966290" watchObservedRunningTime="2025-09-30 14:15:29.241464064 +0000 UTC m=+1157.870550497" Sep 30 14:15:29 crc kubenswrapper[4840]: I0930 14:15:29.241833 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.241827524 podStartE2EDuration="3.241827524s" podCreationTimestamp="2025-09-30 14:15:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:15:28.226888271 +0000 UTC m=+1156.855974694" watchObservedRunningTime="2025-09-30 14:15:29.241827524 +0000 UTC m=+1157.870913947" Sep 30 14:15:29 crc kubenswrapper[4840]: I0930 14:15:29.473707 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Sep 30 14:15:30 crc kubenswrapper[4840]: I0930 14:15:30.232757 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96e1874a-e455-497d-8593-fd9f09bd39e3","Type":"ContainerStarted","Data":"5b9699ada7f63158d4d3d2ffddd2a0d54aecba5b16822a9e851485e07e63154b"} Sep 30 14:15:31 crc kubenswrapper[4840]: I0930 14:15:31.480962 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Sep 30 14:15:31 crc kubenswrapper[4840]: I0930 14:15:31.481762 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Sep 30 14:15:32 crc kubenswrapper[4840]: I0930 14:15:32.251312 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96e1874a-e455-497d-8593-fd9f09bd39e3","Type":"ContainerStarted","Data":"2f8c1329c7d560b967b2aa1f9d0df42a8c33aa2a5f2bfc71f6a07ff364fe1ed6"} Sep 30 14:15:32 crc kubenswrapper[4840]: I0930 14:15:32.288068 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.982862407 podStartE2EDuration="7.288048641s" podCreationTimestamp="2025-09-30 14:15:25 +0000 UTC" firstStartedPulling="2025-09-30 14:15:26.061269956 +0000 UTC m=+1154.690356379" lastFinishedPulling="2025-09-30 14:15:31.36645619 +0000 UTC m=+1159.995542613" observedRunningTime="2025-09-30 14:15:32.278776649 +0000 UTC m=+1160.907863102" watchObservedRunningTime="2025-09-30 14:15:32.288048641 +0000 UTC m=+1160.917135054" Sep 30 14:15:32 crc kubenswrapper[4840]: I0930 14:15:32.497911 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a14c8d91-8aff-4a17-8c9f-ce3839b3dbed" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.195:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 30 14:15:32 crc kubenswrapper[4840]: I0930 14:15:32.498190 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a14c8d91-8aff-4a17-8c9f-ce3839b3dbed" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.195:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 30 14:15:32 crc kubenswrapper[4840]: I0930 14:15:32.581899 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Sep 30 14:15:33 crc kubenswrapper[4840]: I0930 14:15:33.259524 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 30 14:15:36 crc kubenswrapper[4840]: I0930 14:15:36.845148 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 30 14:15:36 crc kubenswrapper[4840]: I0930 14:15:36.845512 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 30 14:15:37 crc kubenswrapper[4840]: I0930 14:15:37.581989 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Sep 30 14:15:37 crc kubenswrapper[4840]: I0930 14:15:37.613229 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Sep 30 14:15:37 crc kubenswrapper[4840]: I0930 14:15:37.927796 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2583ecd5-01f5-4e7d-907b-9dc8a3771273" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.197:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 30 14:15:37 crc kubenswrapper[4840]: I0930 14:15:37.927861 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2583ecd5-01f5-4e7d-907b-9dc8a3771273" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.197:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 30 14:15:38 crc kubenswrapper[4840]: I0930 14:15:38.348346 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Sep 30 14:15:41 crc kubenswrapper[4840]: I0930 14:15:41.486705 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Sep 30 14:15:41 crc kubenswrapper[4840]: I0930 14:15:41.487457 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Sep 30 14:15:41 crc kubenswrapper[4840]: I0930 14:15:41.500180 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Sep 30 14:15:41 crc kubenswrapper[4840]: I0930 14:15:41.500419 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Sep 30 14:15:46 crc kubenswrapper[4840]: I0930 14:15:46.850632 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Sep 30 14:15:46 crc kubenswrapper[4840]: I0930 14:15:46.851334 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Sep 30 14:15:46 crc kubenswrapper[4840]: I0930 14:15:46.851654 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Sep 30 14:15:46 crc kubenswrapper[4840]: I0930 14:15:46.851683 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Sep 30 14:15:46 crc kubenswrapper[4840]: I0930 14:15:46.855889 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Sep 30 14:15:46 crc kubenswrapper[4840]: I0930 14:15:46.856184 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Sep 30 14:15:47 crc kubenswrapper[4840]: I0930 14:15:47.036447 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-mxvtv"] Sep 30 14:15:47 crc kubenswrapper[4840]: I0930 14:15:47.039092 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-mxvtv" Sep 30 14:15:47 crc kubenswrapper[4840]: I0930 14:15:47.068476 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-mxvtv"] Sep 30 14:15:47 crc kubenswrapper[4840]: I0930 14:15:47.156310 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03083423-b558-40aa-afda-c29cef215e79-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-mxvtv\" (UID: \"03083423-b558-40aa-afda-c29cef215e79\") " pod="openstack/dnsmasq-dns-59cf4bdb65-mxvtv" Sep 30 14:15:47 crc kubenswrapper[4840]: I0930 14:15:47.158226 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/03083423-b558-40aa-afda-c29cef215e79-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-mxvtv\" (UID: \"03083423-b558-40aa-afda-c29cef215e79\") " pod="openstack/dnsmasq-dns-59cf4bdb65-mxvtv" Sep 30 14:15:47 crc kubenswrapper[4840]: I0930 14:15:47.158284 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/03083423-b558-40aa-afda-c29cef215e79-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-mxvtv\" (UID: \"03083423-b558-40aa-afda-c29cef215e79\") " pod="openstack/dnsmasq-dns-59cf4bdb65-mxvtv" Sep 30 14:15:47 crc kubenswrapper[4840]: I0930 14:15:47.158373 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/03083423-b558-40aa-afda-c29cef215e79-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-mxvtv\" (UID: \"03083423-b558-40aa-afda-c29cef215e79\") " pod="openstack/dnsmasq-dns-59cf4bdb65-mxvtv" Sep 30 14:15:47 crc kubenswrapper[4840]: I0930 14:15:47.158443 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5kvj\" (UniqueName: \"kubernetes.io/projected/03083423-b558-40aa-afda-c29cef215e79-kube-api-access-j5kvj\") pod \"dnsmasq-dns-59cf4bdb65-mxvtv\" (UID: \"03083423-b558-40aa-afda-c29cef215e79\") " pod="openstack/dnsmasq-dns-59cf4bdb65-mxvtv" Sep 30 14:15:47 crc kubenswrapper[4840]: I0930 14:15:47.158523 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03083423-b558-40aa-afda-c29cef215e79-config\") pod \"dnsmasq-dns-59cf4bdb65-mxvtv\" (UID: \"03083423-b558-40aa-afda-c29cef215e79\") " pod="openstack/dnsmasq-dns-59cf4bdb65-mxvtv" Sep 30 14:15:47 crc kubenswrapper[4840]: I0930 14:15:47.260450 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03083423-b558-40aa-afda-c29cef215e79-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-mxvtv\" (UID: \"03083423-b558-40aa-afda-c29cef215e79\") " pod="openstack/dnsmasq-dns-59cf4bdb65-mxvtv" Sep 30 14:15:47 crc kubenswrapper[4840]: I0930 14:15:47.260500 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/03083423-b558-40aa-afda-c29cef215e79-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-mxvtv\" (UID: \"03083423-b558-40aa-afda-c29cef215e79\") " pod="openstack/dnsmasq-dns-59cf4bdb65-mxvtv" Sep 30 14:15:47 crc kubenswrapper[4840]: I0930 14:15:47.260542 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/03083423-b558-40aa-afda-c29cef215e79-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-mxvtv\" (UID: \"03083423-b558-40aa-afda-c29cef215e79\") " pod="openstack/dnsmasq-dns-59cf4bdb65-mxvtv" Sep 30 14:15:47 crc kubenswrapper[4840]: I0930 14:15:47.260593 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/03083423-b558-40aa-afda-c29cef215e79-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-mxvtv\" (UID: \"03083423-b558-40aa-afda-c29cef215e79\") " pod="openstack/dnsmasq-dns-59cf4bdb65-mxvtv" Sep 30 14:15:47 crc kubenswrapper[4840]: I0930 14:15:47.260636 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5kvj\" (UniqueName: \"kubernetes.io/projected/03083423-b558-40aa-afda-c29cef215e79-kube-api-access-j5kvj\") pod \"dnsmasq-dns-59cf4bdb65-mxvtv\" (UID: \"03083423-b558-40aa-afda-c29cef215e79\") " pod="openstack/dnsmasq-dns-59cf4bdb65-mxvtv" Sep 30 14:15:47 crc kubenswrapper[4840]: I0930 14:15:47.260693 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03083423-b558-40aa-afda-c29cef215e79-config\") pod \"dnsmasq-dns-59cf4bdb65-mxvtv\" (UID: \"03083423-b558-40aa-afda-c29cef215e79\") " pod="openstack/dnsmasq-dns-59cf4bdb65-mxvtv" Sep 30 14:15:47 crc kubenswrapper[4840]: I0930 14:15:47.261829 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/03083423-b558-40aa-afda-c29cef215e79-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-mxvtv\" (UID: \"03083423-b558-40aa-afda-c29cef215e79\") " pod="openstack/dnsmasq-dns-59cf4bdb65-mxvtv" Sep 30 14:15:47 crc kubenswrapper[4840]: I0930 14:15:47.262175 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/03083423-b558-40aa-afda-c29cef215e79-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-mxvtv\" (UID: \"03083423-b558-40aa-afda-c29cef215e79\") " pod="openstack/dnsmasq-dns-59cf4bdb65-mxvtv" Sep 30 14:15:47 crc kubenswrapper[4840]: I0930 14:15:47.262533 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03083423-b558-40aa-afda-c29cef215e79-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-mxvtv\" (UID: \"03083423-b558-40aa-afda-c29cef215e79\") " pod="openstack/dnsmasq-dns-59cf4bdb65-mxvtv" Sep 30 14:15:47 crc kubenswrapper[4840]: I0930 14:15:47.263356 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/03083423-b558-40aa-afda-c29cef215e79-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-mxvtv\" (UID: \"03083423-b558-40aa-afda-c29cef215e79\") " pod="openstack/dnsmasq-dns-59cf4bdb65-mxvtv" Sep 30 14:15:47 crc kubenswrapper[4840]: I0930 14:15:47.263842 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03083423-b558-40aa-afda-c29cef215e79-config\") pod \"dnsmasq-dns-59cf4bdb65-mxvtv\" (UID: \"03083423-b558-40aa-afda-c29cef215e79\") " pod="openstack/dnsmasq-dns-59cf4bdb65-mxvtv" Sep 30 14:15:47 crc kubenswrapper[4840]: I0930 14:15:47.290474 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5kvj\" (UniqueName: \"kubernetes.io/projected/03083423-b558-40aa-afda-c29cef215e79-kube-api-access-j5kvj\") pod \"dnsmasq-dns-59cf4bdb65-mxvtv\" (UID: \"03083423-b558-40aa-afda-c29cef215e79\") " pod="openstack/dnsmasq-dns-59cf4bdb65-mxvtv" Sep 30 14:15:47 crc kubenswrapper[4840]: I0930 14:15:47.418340 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-mxvtv" Sep 30 14:15:47 crc kubenswrapper[4840]: I0930 14:15:47.436328 4840 generic.go:334] "Generic (PLEG): container finished" podID="0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d" containerID="8a34fb015dd622d5f4779489064636d959dba820baf62e3f5f3b9216bbcc2214" exitCode=137 Sep 30 14:15:47 crc kubenswrapper[4840]: I0930 14:15:47.436472 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d","Type":"ContainerDied","Data":"8a34fb015dd622d5f4779489064636d959dba820baf62e3f5f3b9216bbcc2214"} Sep 30 14:15:47 crc kubenswrapper[4840]: I0930 14:15:47.576362 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:15:47 crc kubenswrapper[4840]: I0930 14:15:47.671145 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzrc7\" (UniqueName: \"kubernetes.io/projected/0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d-kube-api-access-vzrc7\") pod \"0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d\" (UID: \"0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d\") " Sep 30 14:15:47 crc kubenswrapper[4840]: I0930 14:15:47.671292 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d-combined-ca-bundle\") pod \"0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d\" (UID: \"0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d\") " Sep 30 14:15:47 crc kubenswrapper[4840]: I0930 14:15:47.671382 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d-config-data\") pod \"0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d\" (UID: \"0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d\") " Sep 30 14:15:47 crc kubenswrapper[4840]: I0930 14:15:47.683038 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d-kube-api-access-vzrc7" (OuterVolumeSpecName: "kube-api-access-vzrc7") pod "0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d" (UID: "0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d"). InnerVolumeSpecName "kube-api-access-vzrc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:15:47 crc kubenswrapper[4840]: I0930 14:15:47.709293 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d" (UID: "0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:15:47 crc kubenswrapper[4840]: I0930 14:15:47.713752 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d-config-data" (OuterVolumeSpecName: "config-data") pod "0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d" (UID: "0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:15:47 crc kubenswrapper[4840]: I0930 14:15:47.773516 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:47 crc kubenswrapper[4840]: I0930 14:15:47.773560 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:47 crc kubenswrapper[4840]: I0930 14:15:47.773570 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzrc7\" (UniqueName: \"kubernetes.io/projected/0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d-kube-api-access-vzrc7\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:47 crc kubenswrapper[4840]: I0930 14:15:47.959394 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-mxvtv"] Sep 30 14:15:47 crc kubenswrapper[4840]: W0930 14:15:47.961589 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod03083423_b558_40aa_afda_c29cef215e79.slice/crio-0dfb69949186453edc5ef7345eecf59f56b9451b05f5fac3ac24e33abf7fce2d WatchSource:0}: Error finding container 0dfb69949186453edc5ef7345eecf59f56b9451b05f5fac3ac24e33abf7fce2d: Status 404 returned error can't find the container with id 0dfb69949186453edc5ef7345eecf59f56b9451b05f5fac3ac24e33abf7fce2d Sep 30 14:15:48 crc kubenswrapper[4840]: I0930 14:15:48.448028 4840 generic.go:334] "Generic (PLEG): container finished" podID="03083423-b558-40aa-afda-c29cef215e79" containerID="952a44017306d2fd9b1f12d571014db74c53629ec6dfbb16338203af55cf8d74" exitCode=0 Sep 30 14:15:48 crc kubenswrapper[4840]: I0930 14:15:48.448173 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-mxvtv" event={"ID":"03083423-b558-40aa-afda-c29cef215e79","Type":"ContainerDied","Data":"952a44017306d2fd9b1f12d571014db74c53629ec6dfbb16338203af55cf8d74"} Sep 30 14:15:48 crc kubenswrapper[4840]: I0930 14:15:48.448989 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-mxvtv" event={"ID":"03083423-b558-40aa-afda-c29cef215e79","Type":"ContainerStarted","Data":"0dfb69949186453edc5ef7345eecf59f56b9451b05f5fac3ac24e33abf7fce2d"} Sep 30 14:15:48 crc kubenswrapper[4840]: I0930 14:15:48.451972 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d","Type":"ContainerDied","Data":"91ac81520ec514e2569ca168378bf0052bc6172ba51c8488f0c4ccf92a2abfeb"} Sep 30 14:15:48 crc kubenswrapper[4840]: I0930 14:15:48.452017 4840 scope.go:117] "RemoveContainer" containerID="8a34fb015dd622d5f4779489064636d959dba820baf62e3f5f3b9216bbcc2214" Sep 30 14:15:48 crc kubenswrapper[4840]: I0930 14:15:48.451981 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:15:48 crc kubenswrapper[4840]: I0930 14:15:48.494624 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 30 14:15:48 crc kubenswrapper[4840]: I0930 14:15:48.504610 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 30 14:15:48 crc kubenswrapper[4840]: I0930 14:15:48.518590 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 30 14:15:48 crc kubenswrapper[4840]: E0930 14:15:48.519092 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d" containerName="nova-cell1-novncproxy-novncproxy" Sep 30 14:15:48 crc kubenswrapper[4840]: I0930 14:15:48.519112 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d" containerName="nova-cell1-novncproxy-novncproxy" Sep 30 14:15:48 crc kubenswrapper[4840]: I0930 14:15:48.519391 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d" containerName="nova-cell1-novncproxy-novncproxy" Sep 30 14:15:48 crc kubenswrapper[4840]: I0930 14:15:48.520176 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:15:48 crc kubenswrapper[4840]: I0930 14:15:48.522942 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Sep 30 14:15:48 crc kubenswrapper[4840]: I0930 14:15:48.523215 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Sep 30 14:15:48 crc kubenswrapper[4840]: I0930 14:15:48.523369 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Sep 30 14:15:48 crc kubenswrapper[4840]: I0930 14:15:48.526700 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 30 14:15:48 crc kubenswrapper[4840]: I0930 14:15:48.589051 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/194d271a-568f-4295-84f2-80cf46e1f123-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"194d271a-568f-4295-84f2-80cf46e1f123\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:15:48 crc kubenswrapper[4840]: I0930 14:15:48.589134 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/194d271a-568f-4295-84f2-80cf46e1f123-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"194d271a-568f-4295-84f2-80cf46e1f123\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:15:48 crc kubenswrapper[4840]: I0930 14:15:48.589217 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lr2xb\" (UniqueName: \"kubernetes.io/projected/194d271a-568f-4295-84f2-80cf46e1f123-kube-api-access-lr2xb\") pod \"nova-cell1-novncproxy-0\" (UID: \"194d271a-568f-4295-84f2-80cf46e1f123\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:15:48 crc kubenswrapper[4840]: I0930 14:15:48.589237 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/194d271a-568f-4295-84f2-80cf46e1f123-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"194d271a-568f-4295-84f2-80cf46e1f123\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:15:48 crc kubenswrapper[4840]: I0930 14:15:48.589282 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/194d271a-568f-4295-84f2-80cf46e1f123-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"194d271a-568f-4295-84f2-80cf46e1f123\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:15:48 crc kubenswrapper[4840]: I0930 14:15:48.691479 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/194d271a-568f-4295-84f2-80cf46e1f123-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"194d271a-568f-4295-84f2-80cf46e1f123\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:15:48 crc kubenswrapper[4840]: I0930 14:15:48.691833 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lr2xb\" (UniqueName: \"kubernetes.io/projected/194d271a-568f-4295-84f2-80cf46e1f123-kube-api-access-lr2xb\") pod \"nova-cell1-novncproxy-0\" (UID: \"194d271a-568f-4295-84f2-80cf46e1f123\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:15:48 crc kubenswrapper[4840]: I0930 14:15:48.691870 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/194d271a-568f-4295-84f2-80cf46e1f123-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"194d271a-568f-4295-84f2-80cf46e1f123\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:15:48 crc kubenswrapper[4840]: I0930 14:15:48.691897 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/194d271a-568f-4295-84f2-80cf46e1f123-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"194d271a-568f-4295-84f2-80cf46e1f123\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:15:48 crc kubenswrapper[4840]: I0930 14:15:48.692452 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/194d271a-568f-4295-84f2-80cf46e1f123-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"194d271a-568f-4295-84f2-80cf46e1f123\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:15:48 crc kubenswrapper[4840]: I0930 14:15:48.700337 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/194d271a-568f-4295-84f2-80cf46e1f123-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"194d271a-568f-4295-84f2-80cf46e1f123\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:15:48 crc kubenswrapper[4840]: I0930 14:15:48.700602 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/194d271a-568f-4295-84f2-80cf46e1f123-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"194d271a-568f-4295-84f2-80cf46e1f123\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:15:48 crc kubenswrapper[4840]: I0930 14:15:48.701100 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/194d271a-568f-4295-84f2-80cf46e1f123-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"194d271a-568f-4295-84f2-80cf46e1f123\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:15:48 crc kubenswrapper[4840]: I0930 14:15:48.701226 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/194d271a-568f-4295-84f2-80cf46e1f123-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"194d271a-568f-4295-84f2-80cf46e1f123\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:15:48 crc kubenswrapper[4840]: I0930 14:15:48.708572 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lr2xb\" (UniqueName: \"kubernetes.io/projected/194d271a-568f-4295-84f2-80cf46e1f123-kube-api-access-lr2xb\") pod \"nova-cell1-novncproxy-0\" (UID: \"194d271a-568f-4295-84f2-80cf46e1f123\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:15:48 crc kubenswrapper[4840]: I0930 14:15:48.945087 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:15:49 crc kubenswrapper[4840]: I0930 14:15:49.233613 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:15:49 crc kubenswrapper[4840]: I0930 14:15:49.234226 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="96e1874a-e455-497d-8593-fd9f09bd39e3" containerName="ceilometer-central-agent" containerID="cri-o://e748188bb7802680970886e08f18bf5277da7e67c46d1ac73a10235551660bb5" gracePeriod=30 Sep 30 14:15:49 crc kubenswrapper[4840]: I0930 14:15:49.234836 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="96e1874a-e455-497d-8593-fd9f09bd39e3" containerName="proxy-httpd" containerID="cri-o://2f8c1329c7d560b967b2aa1f9d0df42a8c33aa2a5f2bfc71f6a07ff364fe1ed6" gracePeriod=30 Sep 30 14:15:49 crc kubenswrapper[4840]: I0930 14:15:49.235535 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="96e1874a-e455-497d-8593-fd9f09bd39e3" containerName="ceilometer-notification-agent" containerID="cri-o://27a04dcb560f3443c2b1c2bf943cac491fbc6a58af74155372c5516866e459ab" gracePeriod=30 Sep 30 14:15:49 crc kubenswrapper[4840]: I0930 14:15:49.235759 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="96e1874a-e455-497d-8593-fd9f09bd39e3" containerName="sg-core" containerID="cri-o://5b9699ada7f63158d4d3d2ffddd2a0d54aecba5b16822a9e851485e07e63154b" gracePeriod=30 Sep 30 14:15:49 crc kubenswrapper[4840]: I0930 14:15:49.340073 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="96e1874a-e455-497d-8593-fd9f09bd39e3" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.196:3000/\": read tcp 10.217.0.2:53956->10.217.0.196:3000: read: connection reset by peer" Sep 30 14:15:49 crc kubenswrapper[4840]: I0930 14:15:49.427347 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 30 14:15:49 crc kubenswrapper[4840]: I0930 14:15:49.469111 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-mxvtv" event={"ID":"03083423-b558-40aa-afda-c29cef215e79","Type":"ContainerStarted","Data":"145a9bfae9f1aa099ac4d9ad79e5402389fe0a558a6fd79bac70812085b8b451"} Sep 30 14:15:49 crc kubenswrapper[4840]: I0930 14:15:49.469177 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59cf4bdb65-mxvtv" Sep 30 14:15:49 crc kubenswrapper[4840]: I0930 14:15:49.473868 4840 generic.go:334] "Generic (PLEG): container finished" podID="96e1874a-e455-497d-8593-fd9f09bd39e3" containerID="2f8c1329c7d560b967b2aa1f9d0df42a8c33aa2a5f2bfc71f6a07ff364fe1ed6" exitCode=0 Sep 30 14:15:49 crc kubenswrapper[4840]: I0930 14:15:49.473893 4840 generic.go:334] "Generic (PLEG): container finished" podID="96e1874a-e455-497d-8593-fd9f09bd39e3" containerID="5b9699ada7f63158d4d3d2ffddd2a0d54aecba5b16822a9e851485e07e63154b" exitCode=2 Sep 30 14:15:49 crc kubenswrapper[4840]: I0930 14:15:49.473933 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96e1874a-e455-497d-8593-fd9f09bd39e3","Type":"ContainerDied","Data":"2f8c1329c7d560b967b2aa1f9d0df42a8c33aa2a5f2bfc71f6a07ff364fe1ed6"} Sep 30 14:15:49 crc kubenswrapper[4840]: I0930 14:15:49.474016 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96e1874a-e455-497d-8593-fd9f09bd39e3","Type":"ContainerDied","Data":"5b9699ada7f63158d4d3d2ffddd2a0d54aecba5b16822a9e851485e07e63154b"} Sep 30 14:15:49 crc kubenswrapper[4840]: I0930 14:15:49.477801 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"194d271a-568f-4295-84f2-80cf46e1f123","Type":"ContainerStarted","Data":"b78a85899b4a4cba71db53b118c558032a3bdd9b2bde0f7fbeb1efcdb0166447"} Sep 30 14:15:49 crc kubenswrapper[4840]: I0930 14:15:49.498074 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59cf4bdb65-mxvtv" podStartSLOduration=2.498052156 podStartE2EDuration="2.498052156s" podCreationTimestamp="2025-09-30 14:15:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:15:49.486290076 +0000 UTC m=+1178.115376499" watchObservedRunningTime="2025-09-30 14:15:49.498052156 +0000 UTC m=+1178.127138579" Sep 30 14:15:49 crc kubenswrapper[4840]: I0930 14:15:49.572670 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 30 14:15:49 crc kubenswrapper[4840]: I0930 14:15:49.572921 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2583ecd5-01f5-4e7d-907b-9dc8a3771273" containerName="nova-api-log" containerID="cri-o://4afc1cadc812e77af8b425f63eb80d4af24749da13a5d82629a2e1d512110e57" gracePeriod=30 Sep 30 14:15:49 crc kubenswrapper[4840]: I0930 14:15:49.573049 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2583ecd5-01f5-4e7d-907b-9dc8a3771273" containerName="nova-api-api" containerID="cri-o://7ab0e5f8e7af49a56cc18f763c4df49e9d070427438a08aff82d02d580adb13b" gracePeriod=30 Sep 30 14:15:50 crc kubenswrapper[4840]: I0930 14:15:50.137115 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d" path="/var/lib/kubelet/pods/0cdc7cab-8e91-4fbe-b80c-1cb23bcbdf6d/volumes" Sep 30 14:15:50 crc kubenswrapper[4840]: I0930 14:15:50.491228 4840 generic.go:334] "Generic (PLEG): container finished" podID="2583ecd5-01f5-4e7d-907b-9dc8a3771273" containerID="4afc1cadc812e77af8b425f63eb80d4af24749da13a5d82629a2e1d512110e57" exitCode=143 Sep 30 14:15:50 crc kubenswrapper[4840]: I0930 14:15:50.491382 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2583ecd5-01f5-4e7d-907b-9dc8a3771273","Type":"ContainerDied","Data":"4afc1cadc812e77af8b425f63eb80d4af24749da13a5d82629a2e1d512110e57"} Sep 30 14:15:50 crc kubenswrapper[4840]: I0930 14:15:50.499006 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"194d271a-568f-4295-84f2-80cf46e1f123","Type":"ContainerStarted","Data":"da3a86039f159daaaff785fe972185a88b589c43a8551f8a5798abb89f0036ea"} Sep 30 14:15:50 crc kubenswrapper[4840]: I0930 14:15:50.513068 4840 generic.go:334] "Generic (PLEG): container finished" podID="96e1874a-e455-497d-8593-fd9f09bd39e3" containerID="27a04dcb560f3443c2b1c2bf943cac491fbc6a58af74155372c5516866e459ab" exitCode=0 Sep 30 14:15:50 crc kubenswrapper[4840]: I0930 14:15:50.513117 4840 generic.go:334] "Generic (PLEG): container finished" podID="96e1874a-e455-497d-8593-fd9f09bd39e3" containerID="e748188bb7802680970886e08f18bf5277da7e67c46d1ac73a10235551660bb5" exitCode=0 Sep 30 14:15:50 crc kubenswrapper[4840]: I0930 14:15:50.513533 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96e1874a-e455-497d-8593-fd9f09bd39e3","Type":"ContainerDied","Data":"27a04dcb560f3443c2b1c2bf943cac491fbc6a58af74155372c5516866e459ab"} Sep 30 14:15:50 crc kubenswrapper[4840]: I0930 14:15:50.513595 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96e1874a-e455-497d-8593-fd9f09bd39e3","Type":"ContainerDied","Data":"e748188bb7802680970886e08f18bf5277da7e67c46d1ac73a10235551660bb5"} Sep 30 14:15:50 crc kubenswrapper[4840]: I0930 14:15:50.523113 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.5230169140000003 podStartE2EDuration="2.523016914s" podCreationTimestamp="2025-09-30 14:15:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:15:50.519671336 +0000 UTC m=+1179.148757759" watchObservedRunningTime="2025-09-30 14:15:50.523016914 +0000 UTC m=+1179.152103337" Sep 30 14:15:50 crc kubenswrapper[4840]: I0930 14:15:50.865752 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:15:50 crc kubenswrapper[4840]: I0930 14:15:50.939236 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96e1874a-e455-497d-8593-fd9f09bd39e3-log-httpd\") pod \"96e1874a-e455-497d-8593-fd9f09bd39e3\" (UID: \"96e1874a-e455-497d-8593-fd9f09bd39e3\") " Sep 30 14:15:50 crc kubenswrapper[4840]: I0930 14:15:50.939311 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4xsz\" (UniqueName: \"kubernetes.io/projected/96e1874a-e455-497d-8593-fd9f09bd39e3-kube-api-access-d4xsz\") pod \"96e1874a-e455-497d-8593-fd9f09bd39e3\" (UID: \"96e1874a-e455-497d-8593-fd9f09bd39e3\") " Sep 30 14:15:50 crc kubenswrapper[4840]: I0930 14:15:50.939348 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/96e1874a-e455-497d-8593-fd9f09bd39e3-ceilometer-tls-certs\") pod \"96e1874a-e455-497d-8593-fd9f09bd39e3\" (UID: \"96e1874a-e455-497d-8593-fd9f09bd39e3\") " Sep 30 14:15:50 crc kubenswrapper[4840]: I0930 14:15:50.939366 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/96e1874a-e455-497d-8593-fd9f09bd39e3-sg-core-conf-yaml\") pod \"96e1874a-e455-497d-8593-fd9f09bd39e3\" (UID: \"96e1874a-e455-497d-8593-fd9f09bd39e3\") " Sep 30 14:15:50 crc kubenswrapper[4840]: I0930 14:15:50.939405 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96e1874a-e455-497d-8593-fd9f09bd39e3-config-data\") pod \"96e1874a-e455-497d-8593-fd9f09bd39e3\" (UID: \"96e1874a-e455-497d-8593-fd9f09bd39e3\") " Sep 30 14:15:50 crc kubenswrapper[4840]: I0930 14:15:50.939492 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96e1874a-e455-497d-8593-fd9f09bd39e3-run-httpd\") pod \"96e1874a-e455-497d-8593-fd9f09bd39e3\" (UID: \"96e1874a-e455-497d-8593-fd9f09bd39e3\") " Sep 30 14:15:50 crc kubenswrapper[4840]: I0930 14:15:50.939522 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96e1874a-e455-497d-8593-fd9f09bd39e3-scripts\") pod \"96e1874a-e455-497d-8593-fd9f09bd39e3\" (UID: \"96e1874a-e455-497d-8593-fd9f09bd39e3\") " Sep 30 14:15:50 crc kubenswrapper[4840]: I0930 14:15:50.939563 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96e1874a-e455-497d-8593-fd9f09bd39e3-combined-ca-bundle\") pod \"96e1874a-e455-497d-8593-fd9f09bd39e3\" (UID: \"96e1874a-e455-497d-8593-fd9f09bd39e3\") " Sep 30 14:15:50 crc kubenswrapper[4840]: I0930 14:15:50.943057 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96e1874a-e455-497d-8593-fd9f09bd39e3-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "96e1874a-e455-497d-8593-fd9f09bd39e3" (UID: "96e1874a-e455-497d-8593-fd9f09bd39e3"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:15:50 crc kubenswrapper[4840]: I0930 14:15:50.944114 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96e1874a-e455-497d-8593-fd9f09bd39e3-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "96e1874a-e455-497d-8593-fd9f09bd39e3" (UID: "96e1874a-e455-497d-8593-fd9f09bd39e3"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:15:50 crc kubenswrapper[4840]: I0930 14:15:50.948524 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96e1874a-e455-497d-8593-fd9f09bd39e3-kube-api-access-d4xsz" (OuterVolumeSpecName: "kube-api-access-d4xsz") pod "96e1874a-e455-497d-8593-fd9f09bd39e3" (UID: "96e1874a-e455-497d-8593-fd9f09bd39e3"). InnerVolumeSpecName "kube-api-access-d4xsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:15:50 crc kubenswrapper[4840]: I0930 14:15:50.969645 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96e1874a-e455-497d-8593-fd9f09bd39e3-scripts" (OuterVolumeSpecName: "scripts") pod "96e1874a-e455-497d-8593-fd9f09bd39e3" (UID: "96e1874a-e455-497d-8593-fd9f09bd39e3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.009802 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96e1874a-e455-497d-8593-fd9f09bd39e3-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "96e1874a-e455-497d-8593-fd9f09bd39e3" (UID: "96e1874a-e455-497d-8593-fd9f09bd39e3"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.016127 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96e1874a-e455-497d-8593-fd9f09bd39e3-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "96e1874a-e455-497d-8593-fd9f09bd39e3" (UID: "96e1874a-e455-497d-8593-fd9f09bd39e3"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.041190 4840 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96e1874a-e455-497d-8593-fd9f09bd39e3-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.041221 4840 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96e1874a-e455-497d-8593-fd9f09bd39e3-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.041229 4840 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96e1874a-e455-497d-8593-fd9f09bd39e3-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.041238 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4xsz\" (UniqueName: \"kubernetes.io/projected/96e1874a-e455-497d-8593-fd9f09bd39e3-kube-api-access-d4xsz\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.041248 4840 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/96e1874a-e455-497d-8593-fd9f09bd39e3-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.041256 4840 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/96e1874a-e455-497d-8593-fd9f09bd39e3-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.054046 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96e1874a-e455-497d-8593-fd9f09bd39e3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "96e1874a-e455-497d-8593-fd9f09bd39e3" (UID: "96e1874a-e455-497d-8593-fd9f09bd39e3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.110764 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96e1874a-e455-497d-8593-fd9f09bd39e3-config-data" (OuterVolumeSpecName: "config-data") pod "96e1874a-e455-497d-8593-fd9f09bd39e3" (UID: "96e1874a-e455-497d-8593-fd9f09bd39e3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.142496 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96e1874a-e455-497d-8593-fd9f09bd39e3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.142535 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96e1874a-e455-497d-8593-fd9f09bd39e3-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.523584 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96e1874a-e455-497d-8593-fd9f09bd39e3","Type":"ContainerDied","Data":"cbb2d50ac7a03cdf6b68437c33714eb624e780cce6bd6e1b9e0496013ad3a9d6"} Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.523890 4840 scope.go:117] "RemoveContainer" containerID="2f8c1329c7d560b967b2aa1f9d0df42a8c33aa2a5f2bfc71f6a07ff364fe1ed6" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.523665 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.560132 4840 scope.go:117] "RemoveContainer" containerID="5b9699ada7f63158d4d3d2ffddd2a0d54aecba5b16822a9e851485e07e63154b" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.592793 4840 scope.go:117] "RemoveContainer" containerID="27a04dcb560f3443c2b1c2bf943cac491fbc6a58af74155372c5516866e459ab" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.595677 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.608444 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.618211 4840 scope.go:117] "RemoveContainer" containerID="e748188bb7802680970886e08f18bf5277da7e67c46d1ac73a10235551660bb5" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.629438 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:15:51 crc kubenswrapper[4840]: E0930 14:15:51.629854 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96e1874a-e455-497d-8593-fd9f09bd39e3" containerName="ceilometer-central-agent" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.629870 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="96e1874a-e455-497d-8593-fd9f09bd39e3" containerName="ceilometer-central-agent" Sep 30 14:15:51 crc kubenswrapper[4840]: E0930 14:15:51.629890 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96e1874a-e455-497d-8593-fd9f09bd39e3" containerName="sg-core" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.629896 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="96e1874a-e455-497d-8593-fd9f09bd39e3" containerName="sg-core" Sep 30 14:15:51 crc kubenswrapper[4840]: E0930 14:15:51.629910 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96e1874a-e455-497d-8593-fd9f09bd39e3" containerName="ceilometer-notification-agent" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.629916 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="96e1874a-e455-497d-8593-fd9f09bd39e3" containerName="ceilometer-notification-agent" Sep 30 14:15:51 crc kubenswrapper[4840]: E0930 14:15:51.629944 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96e1874a-e455-497d-8593-fd9f09bd39e3" containerName="proxy-httpd" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.629952 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="96e1874a-e455-497d-8593-fd9f09bd39e3" containerName="proxy-httpd" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.630166 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="96e1874a-e455-497d-8593-fd9f09bd39e3" containerName="ceilometer-notification-agent" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.630186 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="96e1874a-e455-497d-8593-fd9f09bd39e3" containerName="ceilometer-central-agent" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.630207 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="96e1874a-e455-497d-8593-fd9f09bd39e3" containerName="sg-core" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.630218 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="96e1874a-e455-497d-8593-fd9f09bd39e3" containerName="proxy-httpd" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.631822 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.634426 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.634718 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.664649 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.666131 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.761987 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5673b7b1-d1e1-47ac-8a65-c210430ac4a5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5673b7b1-d1e1-47ac-8a65-c210430ac4a5\") " pod="openstack/ceilometer-0" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.762347 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5673b7b1-d1e1-47ac-8a65-c210430ac4a5-config-data\") pod \"ceilometer-0\" (UID: \"5673b7b1-d1e1-47ac-8a65-c210430ac4a5\") " pod="openstack/ceilometer-0" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.762516 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6v9rg\" (UniqueName: \"kubernetes.io/projected/5673b7b1-d1e1-47ac-8a65-c210430ac4a5-kube-api-access-6v9rg\") pod \"ceilometer-0\" (UID: \"5673b7b1-d1e1-47ac-8a65-c210430ac4a5\") " pod="openstack/ceilometer-0" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.762658 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5673b7b1-d1e1-47ac-8a65-c210430ac4a5-scripts\") pod \"ceilometer-0\" (UID: \"5673b7b1-d1e1-47ac-8a65-c210430ac4a5\") " pod="openstack/ceilometer-0" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.762803 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5673b7b1-d1e1-47ac-8a65-c210430ac4a5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5673b7b1-d1e1-47ac-8a65-c210430ac4a5\") " pod="openstack/ceilometer-0" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.762937 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5673b7b1-d1e1-47ac-8a65-c210430ac4a5-log-httpd\") pod \"ceilometer-0\" (UID: \"5673b7b1-d1e1-47ac-8a65-c210430ac4a5\") " pod="openstack/ceilometer-0" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.763102 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5673b7b1-d1e1-47ac-8a65-c210430ac4a5-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"5673b7b1-d1e1-47ac-8a65-c210430ac4a5\") " pod="openstack/ceilometer-0" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.763267 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5673b7b1-d1e1-47ac-8a65-c210430ac4a5-run-httpd\") pod \"ceilometer-0\" (UID: \"5673b7b1-d1e1-47ac-8a65-c210430ac4a5\") " pod="openstack/ceilometer-0" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.864497 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5673b7b1-d1e1-47ac-8a65-c210430ac4a5-config-data\") pod \"ceilometer-0\" (UID: \"5673b7b1-d1e1-47ac-8a65-c210430ac4a5\") " pod="openstack/ceilometer-0" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.864566 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6v9rg\" (UniqueName: \"kubernetes.io/projected/5673b7b1-d1e1-47ac-8a65-c210430ac4a5-kube-api-access-6v9rg\") pod \"ceilometer-0\" (UID: \"5673b7b1-d1e1-47ac-8a65-c210430ac4a5\") " pod="openstack/ceilometer-0" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.864583 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5673b7b1-d1e1-47ac-8a65-c210430ac4a5-scripts\") pod \"ceilometer-0\" (UID: \"5673b7b1-d1e1-47ac-8a65-c210430ac4a5\") " pod="openstack/ceilometer-0" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.864602 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5673b7b1-d1e1-47ac-8a65-c210430ac4a5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5673b7b1-d1e1-47ac-8a65-c210430ac4a5\") " pod="openstack/ceilometer-0" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.864616 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5673b7b1-d1e1-47ac-8a65-c210430ac4a5-log-httpd\") pod \"ceilometer-0\" (UID: \"5673b7b1-d1e1-47ac-8a65-c210430ac4a5\") " pod="openstack/ceilometer-0" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.864666 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5673b7b1-d1e1-47ac-8a65-c210430ac4a5-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"5673b7b1-d1e1-47ac-8a65-c210430ac4a5\") " pod="openstack/ceilometer-0" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.864696 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5673b7b1-d1e1-47ac-8a65-c210430ac4a5-run-httpd\") pod \"ceilometer-0\" (UID: \"5673b7b1-d1e1-47ac-8a65-c210430ac4a5\") " pod="openstack/ceilometer-0" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.864750 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5673b7b1-d1e1-47ac-8a65-c210430ac4a5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5673b7b1-d1e1-47ac-8a65-c210430ac4a5\") " pod="openstack/ceilometer-0" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.865897 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5673b7b1-d1e1-47ac-8a65-c210430ac4a5-log-httpd\") pod \"ceilometer-0\" (UID: \"5673b7b1-d1e1-47ac-8a65-c210430ac4a5\") " pod="openstack/ceilometer-0" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.866138 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5673b7b1-d1e1-47ac-8a65-c210430ac4a5-run-httpd\") pod \"ceilometer-0\" (UID: \"5673b7b1-d1e1-47ac-8a65-c210430ac4a5\") " pod="openstack/ceilometer-0" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.869943 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5673b7b1-d1e1-47ac-8a65-c210430ac4a5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5673b7b1-d1e1-47ac-8a65-c210430ac4a5\") " pod="openstack/ceilometer-0" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.870726 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5673b7b1-d1e1-47ac-8a65-c210430ac4a5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5673b7b1-d1e1-47ac-8a65-c210430ac4a5\") " pod="openstack/ceilometer-0" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.872078 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5673b7b1-d1e1-47ac-8a65-c210430ac4a5-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"5673b7b1-d1e1-47ac-8a65-c210430ac4a5\") " pod="openstack/ceilometer-0" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.875574 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5673b7b1-d1e1-47ac-8a65-c210430ac4a5-config-data\") pod \"ceilometer-0\" (UID: \"5673b7b1-d1e1-47ac-8a65-c210430ac4a5\") " pod="openstack/ceilometer-0" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.876295 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5673b7b1-d1e1-47ac-8a65-c210430ac4a5-scripts\") pod \"ceilometer-0\" (UID: \"5673b7b1-d1e1-47ac-8a65-c210430ac4a5\") " pod="openstack/ceilometer-0" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.898521 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6v9rg\" (UniqueName: \"kubernetes.io/projected/5673b7b1-d1e1-47ac-8a65-c210430ac4a5-kube-api-access-6v9rg\") pod \"ceilometer-0\" (UID: \"5673b7b1-d1e1-47ac-8a65-c210430ac4a5\") " pod="openstack/ceilometer-0" Sep 30 14:15:51 crc kubenswrapper[4840]: I0930 14:15:51.978651 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:15:52 crc kubenswrapper[4840]: I0930 14:15:52.130843 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96e1874a-e455-497d-8593-fd9f09bd39e3" path="/var/lib/kubelet/pods/96e1874a-e455-497d-8593-fd9f09bd39e3/volumes" Sep 30 14:15:52 crc kubenswrapper[4840]: I0930 14:15:52.404104 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:15:52 crc kubenswrapper[4840]: I0930 14:15:52.534284 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5673b7b1-d1e1-47ac-8a65-c210430ac4a5","Type":"ContainerStarted","Data":"c389124fcc62f228e78eb4a3e87f46ef151fbfeddf5dc1aa09e2834924c47f39"} Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.217954 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.289763 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2583ecd5-01f5-4e7d-907b-9dc8a3771273-config-data\") pod \"2583ecd5-01f5-4e7d-907b-9dc8a3771273\" (UID: \"2583ecd5-01f5-4e7d-907b-9dc8a3771273\") " Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.289833 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2583ecd5-01f5-4e7d-907b-9dc8a3771273-combined-ca-bundle\") pod \"2583ecd5-01f5-4e7d-907b-9dc8a3771273\" (UID: \"2583ecd5-01f5-4e7d-907b-9dc8a3771273\") " Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.289873 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2583ecd5-01f5-4e7d-907b-9dc8a3771273-logs\") pod \"2583ecd5-01f5-4e7d-907b-9dc8a3771273\" (UID: \"2583ecd5-01f5-4e7d-907b-9dc8a3771273\") " Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.290049 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpxxf\" (UniqueName: \"kubernetes.io/projected/2583ecd5-01f5-4e7d-907b-9dc8a3771273-kube-api-access-tpxxf\") pod \"2583ecd5-01f5-4e7d-907b-9dc8a3771273\" (UID: \"2583ecd5-01f5-4e7d-907b-9dc8a3771273\") " Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.291248 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2583ecd5-01f5-4e7d-907b-9dc8a3771273-logs" (OuterVolumeSpecName: "logs") pod "2583ecd5-01f5-4e7d-907b-9dc8a3771273" (UID: "2583ecd5-01f5-4e7d-907b-9dc8a3771273"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.318267 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2583ecd5-01f5-4e7d-907b-9dc8a3771273-kube-api-access-tpxxf" (OuterVolumeSpecName: "kube-api-access-tpxxf") pod "2583ecd5-01f5-4e7d-907b-9dc8a3771273" (UID: "2583ecd5-01f5-4e7d-907b-9dc8a3771273"). InnerVolumeSpecName "kube-api-access-tpxxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.345010 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2583ecd5-01f5-4e7d-907b-9dc8a3771273-config-data" (OuterVolumeSpecName: "config-data") pod "2583ecd5-01f5-4e7d-907b-9dc8a3771273" (UID: "2583ecd5-01f5-4e7d-907b-9dc8a3771273"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.350669 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2583ecd5-01f5-4e7d-907b-9dc8a3771273-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2583ecd5-01f5-4e7d-907b-9dc8a3771273" (UID: "2583ecd5-01f5-4e7d-907b-9dc8a3771273"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.393802 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tpxxf\" (UniqueName: \"kubernetes.io/projected/2583ecd5-01f5-4e7d-907b-9dc8a3771273-kube-api-access-tpxxf\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.394034 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2583ecd5-01f5-4e7d-907b-9dc8a3771273-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.394098 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2583ecd5-01f5-4e7d-907b-9dc8a3771273-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.394166 4840 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2583ecd5-01f5-4e7d-907b-9dc8a3771273-logs\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.542651 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5673b7b1-d1e1-47ac-8a65-c210430ac4a5","Type":"ContainerStarted","Data":"79571a10c48d92df6cc835c6df0fd26adec35c689778e70f61a25152f045430e"} Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.544750 4840 generic.go:334] "Generic (PLEG): container finished" podID="2583ecd5-01f5-4e7d-907b-9dc8a3771273" containerID="7ab0e5f8e7af49a56cc18f763c4df49e9d070427438a08aff82d02d580adb13b" exitCode=0 Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.544799 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2583ecd5-01f5-4e7d-907b-9dc8a3771273","Type":"ContainerDied","Data":"7ab0e5f8e7af49a56cc18f763c4df49e9d070427438a08aff82d02d580adb13b"} Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.544835 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.544864 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2583ecd5-01f5-4e7d-907b-9dc8a3771273","Type":"ContainerDied","Data":"68c87448b25d4f1cdbc875001fc5684a703a033ab1d0778c69a932f4f3ca64a1"} Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.544891 4840 scope.go:117] "RemoveContainer" containerID="7ab0e5f8e7af49a56cc18f763c4df49e9d070427438a08aff82d02d580adb13b" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.569591 4840 scope.go:117] "RemoveContainer" containerID="4afc1cadc812e77af8b425f63eb80d4af24749da13a5d82629a2e1d512110e57" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.588439 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.593687 4840 scope.go:117] "RemoveContainer" containerID="7ab0e5f8e7af49a56cc18f763c4df49e9d070427438a08aff82d02d580adb13b" Sep 30 14:15:53 crc kubenswrapper[4840]: E0930 14:15:53.598070 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ab0e5f8e7af49a56cc18f763c4df49e9d070427438a08aff82d02d580adb13b\": container with ID starting with 7ab0e5f8e7af49a56cc18f763c4df49e9d070427438a08aff82d02d580adb13b not found: ID does not exist" containerID="7ab0e5f8e7af49a56cc18f763c4df49e9d070427438a08aff82d02d580adb13b" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.598118 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ab0e5f8e7af49a56cc18f763c4df49e9d070427438a08aff82d02d580adb13b"} err="failed to get container status \"7ab0e5f8e7af49a56cc18f763c4df49e9d070427438a08aff82d02d580adb13b\": rpc error: code = NotFound desc = could not find container \"7ab0e5f8e7af49a56cc18f763c4df49e9d070427438a08aff82d02d580adb13b\": container with ID starting with 7ab0e5f8e7af49a56cc18f763c4df49e9d070427438a08aff82d02d580adb13b not found: ID does not exist" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.598145 4840 scope.go:117] "RemoveContainer" containerID="4afc1cadc812e77af8b425f63eb80d4af24749da13a5d82629a2e1d512110e57" Sep 30 14:15:53 crc kubenswrapper[4840]: E0930 14:15:53.599688 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4afc1cadc812e77af8b425f63eb80d4af24749da13a5d82629a2e1d512110e57\": container with ID starting with 4afc1cadc812e77af8b425f63eb80d4af24749da13a5d82629a2e1d512110e57 not found: ID does not exist" containerID="4afc1cadc812e77af8b425f63eb80d4af24749da13a5d82629a2e1d512110e57" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.599713 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4afc1cadc812e77af8b425f63eb80d4af24749da13a5d82629a2e1d512110e57"} err="failed to get container status \"4afc1cadc812e77af8b425f63eb80d4af24749da13a5d82629a2e1d512110e57\": rpc error: code = NotFound desc = could not find container \"4afc1cadc812e77af8b425f63eb80d4af24749da13a5d82629a2e1d512110e57\": container with ID starting with 4afc1cadc812e77af8b425f63eb80d4af24749da13a5d82629a2e1d512110e57 not found: ID does not exist" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.602573 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.611220 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Sep 30 14:15:53 crc kubenswrapper[4840]: E0930 14:15:53.611685 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2583ecd5-01f5-4e7d-907b-9dc8a3771273" containerName="nova-api-api" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.611702 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="2583ecd5-01f5-4e7d-907b-9dc8a3771273" containerName="nova-api-api" Sep 30 14:15:53 crc kubenswrapper[4840]: E0930 14:15:53.611717 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2583ecd5-01f5-4e7d-907b-9dc8a3771273" containerName="nova-api-log" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.611725 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="2583ecd5-01f5-4e7d-907b-9dc8a3771273" containerName="nova-api-log" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.611897 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="2583ecd5-01f5-4e7d-907b-9dc8a3771273" containerName="nova-api-api" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.611916 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="2583ecd5-01f5-4e7d-907b-9dc8a3771273" containerName="nova-api-log" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.620100 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.622545 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.627686 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.628046 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.628137 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.701345 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d816b4be-b475-4a6d-b7c7-e955741fcbba-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d816b4be-b475-4a6d-b7c7-e955741fcbba\") " pod="openstack/nova-api-0" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.701392 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d816b4be-b475-4a6d-b7c7-e955741fcbba-logs\") pod \"nova-api-0\" (UID: \"d816b4be-b475-4a6d-b7c7-e955741fcbba\") " pod="openstack/nova-api-0" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.701417 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d816b4be-b475-4a6d-b7c7-e955741fcbba-public-tls-certs\") pod \"nova-api-0\" (UID: \"d816b4be-b475-4a6d-b7c7-e955741fcbba\") " pod="openstack/nova-api-0" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.701522 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsgjn\" (UniqueName: \"kubernetes.io/projected/d816b4be-b475-4a6d-b7c7-e955741fcbba-kube-api-access-lsgjn\") pod \"nova-api-0\" (UID: \"d816b4be-b475-4a6d-b7c7-e955741fcbba\") " pod="openstack/nova-api-0" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.701562 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d816b4be-b475-4a6d-b7c7-e955741fcbba-config-data\") pod \"nova-api-0\" (UID: \"d816b4be-b475-4a6d-b7c7-e955741fcbba\") " pod="openstack/nova-api-0" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.701755 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d816b4be-b475-4a6d-b7c7-e955741fcbba-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d816b4be-b475-4a6d-b7c7-e955741fcbba\") " pod="openstack/nova-api-0" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.802997 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsgjn\" (UniqueName: \"kubernetes.io/projected/d816b4be-b475-4a6d-b7c7-e955741fcbba-kube-api-access-lsgjn\") pod \"nova-api-0\" (UID: \"d816b4be-b475-4a6d-b7c7-e955741fcbba\") " pod="openstack/nova-api-0" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.803386 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d816b4be-b475-4a6d-b7c7-e955741fcbba-config-data\") pod \"nova-api-0\" (UID: \"d816b4be-b475-4a6d-b7c7-e955741fcbba\") " pod="openstack/nova-api-0" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.804105 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d816b4be-b475-4a6d-b7c7-e955741fcbba-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d816b4be-b475-4a6d-b7c7-e955741fcbba\") " pod="openstack/nova-api-0" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.804343 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d816b4be-b475-4a6d-b7c7-e955741fcbba-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d816b4be-b475-4a6d-b7c7-e955741fcbba\") " pod="openstack/nova-api-0" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.804370 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d816b4be-b475-4a6d-b7c7-e955741fcbba-logs\") pod \"nova-api-0\" (UID: \"d816b4be-b475-4a6d-b7c7-e955741fcbba\") " pod="openstack/nova-api-0" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.804399 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d816b4be-b475-4a6d-b7c7-e955741fcbba-public-tls-certs\") pod \"nova-api-0\" (UID: \"d816b4be-b475-4a6d-b7c7-e955741fcbba\") " pod="openstack/nova-api-0" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.806071 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d816b4be-b475-4a6d-b7c7-e955741fcbba-logs\") pod \"nova-api-0\" (UID: \"d816b4be-b475-4a6d-b7c7-e955741fcbba\") " pod="openstack/nova-api-0" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.808463 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d816b4be-b475-4a6d-b7c7-e955741fcbba-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d816b4be-b475-4a6d-b7c7-e955741fcbba\") " pod="openstack/nova-api-0" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.809223 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d816b4be-b475-4a6d-b7c7-e955741fcbba-config-data\") pod \"nova-api-0\" (UID: \"d816b4be-b475-4a6d-b7c7-e955741fcbba\") " pod="openstack/nova-api-0" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.817860 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d816b4be-b475-4a6d-b7c7-e955741fcbba-public-tls-certs\") pod \"nova-api-0\" (UID: \"d816b4be-b475-4a6d-b7c7-e955741fcbba\") " pod="openstack/nova-api-0" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.820127 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d816b4be-b475-4a6d-b7c7-e955741fcbba-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d816b4be-b475-4a6d-b7c7-e955741fcbba\") " pod="openstack/nova-api-0" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.830593 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsgjn\" (UniqueName: \"kubernetes.io/projected/d816b4be-b475-4a6d-b7c7-e955741fcbba-kube-api-access-lsgjn\") pod \"nova-api-0\" (UID: \"d816b4be-b475-4a6d-b7c7-e955741fcbba\") " pod="openstack/nova-api-0" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.946023 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:15:53 crc kubenswrapper[4840]: I0930 14:15:53.959961 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 14:15:54 crc kubenswrapper[4840]: I0930 14:15:54.134337 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2583ecd5-01f5-4e7d-907b-9dc8a3771273" path="/var/lib/kubelet/pods/2583ecd5-01f5-4e7d-907b-9dc8a3771273/volumes" Sep 30 14:15:54 crc kubenswrapper[4840]: I0930 14:15:54.404853 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 30 14:15:54 crc kubenswrapper[4840]: W0930 14:15:54.414520 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd816b4be_b475_4a6d_b7c7_e955741fcbba.slice/crio-05933c66dad72a17cf2b7c5fa4ac97cf4be31c73690ef1811294e80c7e32499f WatchSource:0}: Error finding container 05933c66dad72a17cf2b7c5fa4ac97cf4be31c73690ef1811294e80c7e32499f: Status 404 returned error can't find the container with id 05933c66dad72a17cf2b7c5fa4ac97cf4be31c73690ef1811294e80c7e32499f Sep 30 14:15:54 crc kubenswrapper[4840]: I0930 14:15:54.555545 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d816b4be-b475-4a6d-b7c7-e955741fcbba","Type":"ContainerStarted","Data":"05933c66dad72a17cf2b7c5fa4ac97cf4be31c73690ef1811294e80c7e32499f"} Sep 30 14:15:55 crc kubenswrapper[4840]: I0930 14:15:55.569887 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d816b4be-b475-4a6d-b7c7-e955741fcbba","Type":"ContainerStarted","Data":"6a13fa59195a1222469f537a473de6efef9f2bfdeb3434d22401e1c2abc9e2b6"} Sep 30 14:15:55 crc kubenswrapper[4840]: I0930 14:15:55.577947 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5673b7b1-d1e1-47ac-8a65-c210430ac4a5","Type":"ContainerStarted","Data":"144ef62f5999bb72670bb41e492912239fd17b93338a6eabb8a18ea34eb76d61"} Sep 30 14:15:56 crc kubenswrapper[4840]: I0930 14:15:56.587929 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d816b4be-b475-4a6d-b7c7-e955741fcbba","Type":"ContainerStarted","Data":"1f59be2c7ab21244947c4b598d95af9dff7a71a384e668a7d6d375e330d18043"} Sep 30 14:15:56 crc kubenswrapper[4840]: I0930 14:15:56.605771 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.605756545 podStartE2EDuration="3.605756545s" podCreationTimestamp="2025-09-30 14:15:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:15:56.605484528 +0000 UTC m=+1185.234570951" watchObservedRunningTime="2025-09-30 14:15:56.605756545 +0000 UTC m=+1185.234842958" Sep 30 14:15:57 crc kubenswrapper[4840]: I0930 14:15:57.420693 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59cf4bdb65-mxvtv" Sep 30 14:15:57 crc kubenswrapper[4840]: I0930 14:15:57.477971 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-kgxlg"] Sep 30 14:15:57 crc kubenswrapper[4840]: I0930 14:15:57.478218 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-845d6d6f59-kgxlg" podUID="85b94340-ea1f-4cde-b359-d4970084e6f5" containerName="dnsmasq-dns" containerID="cri-o://9944d1240049ca763bd1030a91e7bfc104f71e03553fc6b95c95cf3a385e5784" gracePeriod=10 Sep 30 14:15:58 crc kubenswrapper[4840]: I0930 14:15:58.613887 4840 generic.go:334] "Generic (PLEG): container finished" podID="85b94340-ea1f-4cde-b359-d4970084e6f5" containerID="9944d1240049ca763bd1030a91e7bfc104f71e03553fc6b95c95cf3a385e5784" exitCode=0 Sep 30 14:15:58 crc kubenswrapper[4840]: I0930 14:15:58.613950 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-kgxlg" event={"ID":"85b94340-ea1f-4cde-b359-d4970084e6f5","Type":"ContainerDied","Data":"9944d1240049ca763bd1030a91e7bfc104f71e03553fc6b95c95cf3a385e5784"} Sep 30 14:15:58 crc kubenswrapper[4840]: I0930 14:15:58.614493 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-kgxlg" event={"ID":"85b94340-ea1f-4cde-b359-d4970084e6f5","Type":"ContainerDied","Data":"12adbd12bf25d9bc20e7f291061be4abaa64109c9ce2809df8a0801723c1fe00"} Sep 30 14:15:58 crc kubenswrapper[4840]: I0930 14:15:58.614508 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="12adbd12bf25d9bc20e7f291061be4abaa64109c9ce2809df8a0801723c1fe00" Sep 30 14:15:58 crc kubenswrapper[4840]: I0930 14:15:58.617494 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5673b7b1-d1e1-47ac-8a65-c210430ac4a5","Type":"ContainerStarted","Data":"ff902b1e611d6223bbb36f434d54474c385da6bec5815f7b053b3916e1f4bb1a"} Sep 30 14:15:58 crc kubenswrapper[4840]: I0930 14:15:58.681886 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-kgxlg" Sep 30 14:15:58 crc kubenswrapper[4840]: I0930 14:15:58.799215 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/85b94340-ea1f-4cde-b359-d4970084e6f5-ovsdbserver-nb\") pod \"85b94340-ea1f-4cde-b359-d4970084e6f5\" (UID: \"85b94340-ea1f-4cde-b359-d4970084e6f5\") " Sep 30 14:15:58 crc kubenswrapper[4840]: I0930 14:15:58.799262 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/85b94340-ea1f-4cde-b359-d4970084e6f5-ovsdbserver-sb\") pod \"85b94340-ea1f-4cde-b359-d4970084e6f5\" (UID: \"85b94340-ea1f-4cde-b359-d4970084e6f5\") " Sep 30 14:15:58 crc kubenswrapper[4840]: I0930 14:15:58.799346 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/85b94340-ea1f-4cde-b359-d4970084e6f5-dns-swift-storage-0\") pod \"85b94340-ea1f-4cde-b359-d4970084e6f5\" (UID: \"85b94340-ea1f-4cde-b359-d4970084e6f5\") " Sep 30 14:15:58 crc kubenswrapper[4840]: I0930 14:15:58.799424 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqxjp\" (UniqueName: \"kubernetes.io/projected/85b94340-ea1f-4cde-b359-d4970084e6f5-kube-api-access-rqxjp\") pod \"85b94340-ea1f-4cde-b359-d4970084e6f5\" (UID: \"85b94340-ea1f-4cde-b359-d4970084e6f5\") " Sep 30 14:15:58 crc kubenswrapper[4840]: I0930 14:15:58.799448 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/85b94340-ea1f-4cde-b359-d4970084e6f5-dns-svc\") pod \"85b94340-ea1f-4cde-b359-d4970084e6f5\" (UID: \"85b94340-ea1f-4cde-b359-d4970084e6f5\") " Sep 30 14:15:58 crc kubenswrapper[4840]: I0930 14:15:58.799489 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85b94340-ea1f-4cde-b359-d4970084e6f5-config\") pod \"85b94340-ea1f-4cde-b359-d4970084e6f5\" (UID: \"85b94340-ea1f-4cde-b359-d4970084e6f5\") " Sep 30 14:15:58 crc kubenswrapper[4840]: I0930 14:15:58.850877 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85b94340-ea1f-4cde-b359-d4970084e6f5-kube-api-access-rqxjp" (OuterVolumeSpecName: "kube-api-access-rqxjp") pod "85b94340-ea1f-4cde-b359-d4970084e6f5" (UID: "85b94340-ea1f-4cde-b359-d4970084e6f5"). InnerVolumeSpecName "kube-api-access-rqxjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:15:58 crc kubenswrapper[4840]: I0930 14:15:58.902282 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqxjp\" (UniqueName: \"kubernetes.io/projected/85b94340-ea1f-4cde-b359-d4970084e6f5-kube-api-access-rqxjp\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:58 crc kubenswrapper[4840]: I0930 14:15:58.924285 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85b94340-ea1f-4cde-b359-d4970084e6f5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "85b94340-ea1f-4cde-b359-d4970084e6f5" (UID: "85b94340-ea1f-4cde-b359-d4970084e6f5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:15:58 crc kubenswrapper[4840]: I0930 14:15:58.927379 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85b94340-ea1f-4cde-b359-d4970084e6f5-config" (OuterVolumeSpecName: "config") pod "85b94340-ea1f-4cde-b359-d4970084e6f5" (UID: "85b94340-ea1f-4cde-b359-d4970084e6f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:15:58 crc kubenswrapper[4840]: I0930 14:15:58.931167 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85b94340-ea1f-4cde-b359-d4970084e6f5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "85b94340-ea1f-4cde-b359-d4970084e6f5" (UID: "85b94340-ea1f-4cde-b359-d4970084e6f5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:15:58 crc kubenswrapper[4840]: I0930 14:15:58.937397 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85b94340-ea1f-4cde-b359-d4970084e6f5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "85b94340-ea1f-4cde-b359-d4970084e6f5" (UID: "85b94340-ea1f-4cde-b359-d4970084e6f5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:15:58 crc kubenswrapper[4840]: I0930 14:15:58.945993 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:15:58 crc kubenswrapper[4840]: I0930 14:15:58.953119 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85b94340-ea1f-4cde-b359-d4970084e6f5-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "85b94340-ea1f-4cde-b359-d4970084e6f5" (UID: "85b94340-ea1f-4cde-b359-d4970084e6f5"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:15:58 crc kubenswrapper[4840]: I0930 14:15:58.963464 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:15:59 crc kubenswrapper[4840]: I0930 14:15:59.005706 4840 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/85b94340-ea1f-4cde-b359-d4970084e6f5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:59 crc kubenswrapper[4840]: I0930 14:15:59.006401 4840 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/85b94340-ea1f-4cde-b359-d4970084e6f5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:59 crc kubenswrapper[4840]: I0930 14:15:59.006432 4840 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/85b94340-ea1f-4cde-b359-d4970084e6f5-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:59 crc kubenswrapper[4840]: I0930 14:15:59.006447 4840 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/85b94340-ea1f-4cde-b359-d4970084e6f5-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:59 crc kubenswrapper[4840]: I0930 14:15:59.006458 4840 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85b94340-ea1f-4cde-b359-d4970084e6f5-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:15:59 crc kubenswrapper[4840]: I0930 14:15:59.627269 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-kgxlg" Sep 30 14:15:59 crc kubenswrapper[4840]: I0930 14:15:59.642514 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:15:59 crc kubenswrapper[4840]: I0930 14:15:59.681784 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-kgxlg"] Sep 30 14:15:59 crc kubenswrapper[4840]: I0930 14:15:59.695605 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-kgxlg"] Sep 30 14:16:00 crc kubenswrapper[4840]: I0930 14:16:00.138037 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85b94340-ea1f-4cde-b359-d4970084e6f5" path="/var/lib/kubelet/pods/85b94340-ea1f-4cde-b359-d4970084e6f5/volumes" Sep 30 14:16:01 crc kubenswrapper[4840]: I0930 14:16:01.650714 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5673b7b1-d1e1-47ac-8a65-c210430ac4a5","Type":"ContainerStarted","Data":"7c4bee183a93fabac49ea548efa932f28673b9b7ccef68dc6b6c48d7970d4392"} Sep 30 14:16:01 crc kubenswrapper[4840]: I0930 14:16:01.651336 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 30 14:16:01 crc kubenswrapper[4840]: I0930 14:16:01.693432 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.477828249 podStartE2EDuration="10.693392964s" podCreationTimestamp="2025-09-30 14:15:51 +0000 UTC" firstStartedPulling="2025-09-30 14:15:52.413975245 +0000 UTC m=+1181.043061668" lastFinishedPulling="2025-09-30 14:16:00.62953996 +0000 UTC m=+1189.258626383" observedRunningTime="2025-09-30 14:16:01.692023728 +0000 UTC m=+1190.321110171" watchObservedRunningTime="2025-09-30 14:16:01.693392964 +0000 UTC m=+1190.322479387" Sep 30 14:16:03 crc kubenswrapper[4840]: I0930 14:16:03.425394 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-845d6d6f59-kgxlg" podUID="85b94340-ea1f-4cde-b359-d4970084e6f5" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.191:5353: i/o timeout" Sep 30 14:16:03 crc kubenswrapper[4840]: I0930 14:16:03.960691 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 30 14:16:03 crc kubenswrapper[4840]: I0930 14:16:03.960741 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 30 14:16:04 crc kubenswrapper[4840]: I0930 14:16:04.973772 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d816b4be-b475-4a6d-b7c7-e955741fcbba" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.202:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 30 14:16:04 crc kubenswrapper[4840]: I0930 14:16:04.973773 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d816b4be-b475-4a6d-b7c7-e955741fcbba" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.202:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 30 14:16:08 crc kubenswrapper[4840]: I0930 14:16:08.712441 4840 generic.go:334] "Generic (PLEG): container finished" podID="91fc410c-6ef8-4864-8ba2-a604c7bcd8ac" containerID="d2a2e49e9d9bfd23804ae23d4fe6675ebfe92252517fe6c60b4fc629f086bff7" exitCode=0 Sep 30 14:16:08 crc kubenswrapper[4840]: I0930 14:16:08.712488 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-2ph7k" event={"ID":"91fc410c-6ef8-4864-8ba2-a604c7bcd8ac","Type":"ContainerDied","Data":"d2a2e49e9d9bfd23804ae23d4fe6675ebfe92252517fe6c60b4fc629f086bff7"} Sep 30 14:16:10 crc kubenswrapper[4840]: I0930 14:16:10.077702 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-2ph7k" Sep 30 14:16:10 crc kubenswrapper[4840]: I0930 14:16:10.208094 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91fc410c-6ef8-4864-8ba2-a604c7bcd8ac-config-data\") pod \"91fc410c-6ef8-4864-8ba2-a604c7bcd8ac\" (UID: \"91fc410c-6ef8-4864-8ba2-a604c7bcd8ac\") " Sep 30 14:16:10 crc kubenswrapper[4840]: I0930 14:16:10.208492 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hkjq\" (UniqueName: \"kubernetes.io/projected/91fc410c-6ef8-4864-8ba2-a604c7bcd8ac-kube-api-access-2hkjq\") pod \"91fc410c-6ef8-4864-8ba2-a604c7bcd8ac\" (UID: \"91fc410c-6ef8-4864-8ba2-a604c7bcd8ac\") " Sep 30 14:16:10 crc kubenswrapper[4840]: I0930 14:16:10.208525 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91fc410c-6ef8-4864-8ba2-a604c7bcd8ac-scripts\") pod \"91fc410c-6ef8-4864-8ba2-a604c7bcd8ac\" (UID: \"91fc410c-6ef8-4864-8ba2-a604c7bcd8ac\") " Sep 30 14:16:10 crc kubenswrapper[4840]: I0930 14:16:10.208565 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91fc410c-6ef8-4864-8ba2-a604c7bcd8ac-combined-ca-bundle\") pod \"91fc410c-6ef8-4864-8ba2-a604c7bcd8ac\" (UID: \"91fc410c-6ef8-4864-8ba2-a604c7bcd8ac\") " Sep 30 14:16:10 crc kubenswrapper[4840]: I0930 14:16:10.214558 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91fc410c-6ef8-4864-8ba2-a604c7bcd8ac-scripts" (OuterVolumeSpecName: "scripts") pod "91fc410c-6ef8-4864-8ba2-a604c7bcd8ac" (UID: "91fc410c-6ef8-4864-8ba2-a604c7bcd8ac"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:16:10 crc kubenswrapper[4840]: I0930 14:16:10.214576 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91fc410c-6ef8-4864-8ba2-a604c7bcd8ac-kube-api-access-2hkjq" (OuterVolumeSpecName: "kube-api-access-2hkjq") pod "91fc410c-6ef8-4864-8ba2-a604c7bcd8ac" (UID: "91fc410c-6ef8-4864-8ba2-a604c7bcd8ac"). InnerVolumeSpecName "kube-api-access-2hkjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:16:10 crc kubenswrapper[4840]: I0930 14:16:10.243373 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91fc410c-6ef8-4864-8ba2-a604c7bcd8ac-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "91fc410c-6ef8-4864-8ba2-a604c7bcd8ac" (UID: "91fc410c-6ef8-4864-8ba2-a604c7bcd8ac"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:16:10 crc kubenswrapper[4840]: I0930 14:16:10.246361 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91fc410c-6ef8-4864-8ba2-a604c7bcd8ac-config-data" (OuterVolumeSpecName: "config-data") pod "91fc410c-6ef8-4864-8ba2-a604c7bcd8ac" (UID: "91fc410c-6ef8-4864-8ba2-a604c7bcd8ac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:16:10 crc kubenswrapper[4840]: I0930 14:16:10.311446 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91fc410c-6ef8-4864-8ba2-a604c7bcd8ac-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:10 crc kubenswrapper[4840]: I0930 14:16:10.311525 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hkjq\" (UniqueName: \"kubernetes.io/projected/91fc410c-6ef8-4864-8ba2-a604c7bcd8ac-kube-api-access-2hkjq\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:10 crc kubenswrapper[4840]: I0930 14:16:10.311539 4840 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91fc410c-6ef8-4864-8ba2-a604c7bcd8ac-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:10 crc kubenswrapper[4840]: I0930 14:16:10.311567 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91fc410c-6ef8-4864-8ba2-a604c7bcd8ac-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:10 crc kubenswrapper[4840]: I0930 14:16:10.732488 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-2ph7k" event={"ID":"91fc410c-6ef8-4864-8ba2-a604c7bcd8ac","Type":"ContainerDied","Data":"2b27344500ffc8121defe005851d508edfe9d8afabbcbe19cec1a9a44ef6dbd7"} Sep 30 14:16:10 crc kubenswrapper[4840]: I0930 14:16:10.732581 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b27344500ffc8121defe005851d508edfe9d8afabbcbe19cec1a9a44ef6dbd7" Sep 30 14:16:10 crc kubenswrapper[4840]: I0930 14:16:10.732536 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-2ph7k" Sep 30 14:16:10 crc kubenswrapper[4840]: I0930 14:16:10.793005 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Sep 30 14:16:10 crc kubenswrapper[4840]: E0930 14:16:10.793841 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85b94340-ea1f-4cde-b359-d4970084e6f5" containerName="dnsmasq-dns" Sep 30 14:16:10 crc kubenswrapper[4840]: I0930 14:16:10.793971 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="85b94340-ea1f-4cde-b359-d4970084e6f5" containerName="dnsmasq-dns" Sep 30 14:16:10 crc kubenswrapper[4840]: E0930 14:16:10.794055 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85b94340-ea1f-4cde-b359-d4970084e6f5" containerName="init" Sep 30 14:16:10 crc kubenswrapper[4840]: I0930 14:16:10.794121 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="85b94340-ea1f-4cde-b359-d4970084e6f5" containerName="init" Sep 30 14:16:10 crc kubenswrapper[4840]: E0930 14:16:10.794203 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91fc410c-6ef8-4864-8ba2-a604c7bcd8ac" containerName="nova-cell1-conductor-db-sync" Sep 30 14:16:10 crc kubenswrapper[4840]: I0930 14:16:10.794276 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="91fc410c-6ef8-4864-8ba2-a604c7bcd8ac" containerName="nova-cell1-conductor-db-sync" Sep 30 14:16:10 crc kubenswrapper[4840]: I0930 14:16:10.795425 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="85b94340-ea1f-4cde-b359-d4970084e6f5" containerName="dnsmasq-dns" Sep 30 14:16:10 crc kubenswrapper[4840]: I0930 14:16:10.795588 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="91fc410c-6ef8-4864-8ba2-a604c7bcd8ac" containerName="nova-cell1-conductor-db-sync" Sep 30 14:16:10 crc kubenswrapper[4840]: I0930 14:16:10.796476 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Sep 30 14:16:10 crc kubenswrapper[4840]: I0930 14:16:10.800999 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Sep 30 14:16:10 crc kubenswrapper[4840]: I0930 14:16:10.805809 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Sep 30 14:16:10 crc kubenswrapper[4840]: I0930 14:16:10.921645 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9nrb\" (UniqueName: \"kubernetes.io/projected/b63a8a4d-0f06-4f78-9f54-9b4e81183705-kube-api-access-r9nrb\") pod \"nova-cell1-conductor-0\" (UID: \"b63a8a4d-0f06-4f78-9f54-9b4e81183705\") " pod="openstack/nova-cell1-conductor-0" Sep 30 14:16:10 crc kubenswrapper[4840]: I0930 14:16:10.921991 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b63a8a4d-0f06-4f78-9f54-9b4e81183705-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"b63a8a4d-0f06-4f78-9f54-9b4e81183705\") " pod="openstack/nova-cell1-conductor-0" Sep 30 14:16:10 crc kubenswrapper[4840]: I0930 14:16:10.922396 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b63a8a4d-0f06-4f78-9f54-9b4e81183705-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"b63a8a4d-0f06-4f78-9f54-9b4e81183705\") " pod="openstack/nova-cell1-conductor-0" Sep 30 14:16:11 crc kubenswrapper[4840]: I0930 14:16:11.024651 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b63a8a4d-0f06-4f78-9f54-9b4e81183705-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"b63a8a4d-0f06-4f78-9f54-9b4e81183705\") " pod="openstack/nova-cell1-conductor-0" Sep 30 14:16:11 crc kubenswrapper[4840]: I0930 14:16:11.024789 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9nrb\" (UniqueName: \"kubernetes.io/projected/b63a8a4d-0f06-4f78-9f54-9b4e81183705-kube-api-access-r9nrb\") pod \"nova-cell1-conductor-0\" (UID: \"b63a8a4d-0f06-4f78-9f54-9b4e81183705\") " pod="openstack/nova-cell1-conductor-0" Sep 30 14:16:11 crc kubenswrapper[4840]: I0930 14:16:11.024836 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b63a8a4d-0f06-4f78-9f54-9b4e81183705-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"b63a8a4d-0f06-4f78-9f54-9b4e81183705\") " pod="openstack/nova-cell1-conductor-0" Sep 30 14:16:11 crc kubenswrapper[4840]: I0930 14:16:11.030901 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b63a8a4d-0f06-4f78-9f54-9b4e81183705-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"b63a8a4d-0f06-4f78-9f54-9b4e81183705\") " pod="openstack/nova-cell1-conductor-0" Sep 30 14:16:11 crc kubenswrapper[4840]: I0930 14:16:11.031749 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b63a8a4d-0f06-4f78-9f54-9b4e81183705-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"b63a8a4d-0f06-4f78-9f54-9b4e81183705\") " pod="openstack/nova-cell1-conductor-0" Sep 30 14:16:11 crc kubenswrapper[4840]: I0930 14:16:11.042125 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9nrb\" (UniqueName: \"kubernetes.io/projected/b63a8a4d-0f06-4f78-9f54-9b4e81183705-kube-api-access-r9nrb\") pod \"nova-cell1-conductor-0\" (UID: \"b63a8a4d-0f06-4f78-9f54-9b4e81183705\") " pod="openstack/nova-cell1-conductor-0" Sep 30 14:16:11 crc kubenswrapper[4840]: I0930 14:16:11.114075 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Sep 30 14:16:11 crc kubenswrapper[4840]: I0930 14:16:11.540188 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Sep 30 14:16:11 crc kubenswrapper[4840]: W0930 14:16:11.543759 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb63a8a4d_0f06_4f78_9f54_9b4e81183705.slice/crio-b4129c493a22e61dacd1ef2074c2281877592ed2a850117a48cffbb4c27fab9c WatchSource:0}: Error finding container b4129c493a22e61dacd1ef2074c2281877592ed2a850117a48cffbb4c27fab9c: Status 404 returned error can't find the container with id b4129c493a22e61dacd1ef2074c2281877592ed2a850117a48cffbb4c27fab9c Sep 30 14:16:11 crc kubenswrapper[4840]: I0930 14:16:11.742082 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"b63a8a4d-0f06-4f78-9f54-9b4e81183705","Type":"ContainerStarted","Data":"b4129c493a22e61dacd1ef2074c2281877592ed2a850117a48cffbb4c27fab9c"} Sep 30 14:16:12 crc kubenswrapper[4840]: I0930 14:16:12.768785 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"b63a8a4d-0f06-4f78-9f54-9b4e81183705","Type":"ContainerStarted","Data":"18bd816c211c601fa9cc18cc42e2d6203f89de9e88252a494d54b831c735b5bc"} Sep 30 14:16:12 crc kubenswrapper[4840]: I0930 14:16:12.770198 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Sep 30 14:16:12 crc kubenswrapper[4840]: I0930 14:16:12.801226 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.801205854 podStartE2EDuration="2.801205854s" podCreationTimestamp="2025-09-30 14:16:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:16:12.781092503 +0000 UTC m=+1201.410178946" watchObservedRunningTime="2025-09-30 14:16:12.801205854 +0000 UTC m=+1201.430292277" Sep 30 14:16:13 crc kubenswrapper[4840]: I0930 14:16:13.968758 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Sep 30 14:16:13 crc kubenswrapper[4840]: I0930 14:16:13.969799 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Sep 30 14:16:13 crc kubenswrapper[4840]: I0930 14:16:13.970116 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Sep 30 14:16:13 crc kubenswrapper[4840]: I0930 14:16:13.970179 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Sep 30 14:16:13 crc kubenswrapper[4840]: I0930 14:16:13.978214 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Sep 30 14:16:13 crc kubenswrapper[4840]: I0930 14:16:13.980199 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Sep 30 14:16:16 crc kubenswrapper[4840]: I0930 14:16:16.149613 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Sep 30 14:16:17 crc kubenswrapper[4840]: I0930 14:16:17.196760 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-f9pkg"] Sep 30 14:16:17 crc kubenswrapper[4840]: I0930 14:16:17.198753 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-f9pkg" Sep 30 14:16:17 crc kubenswrapper[4840]: I0930 14:16:17.201267 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Sep 30 14:16:17 crc kubenswrapper[4840]: I0930 14:16:17.201302 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Sep 30 14:16:17 crc kubenswrapper[4840]: I0930 14:16:17.217174 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-f9pkg"] Sep 30 14:16:17 crc kubenswrapper[4840]: I0930 14:16:17.358789 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/521c6360-ed9d-4b6d-a74b-188e3a27be00-scripts\") pod \"nova-cell1-cell-mapping-f9pkg\" (UID: \"521c6360-ed9d-4b6d-a74b-188e3a27be00\") " pod="openstack/nova-cell1-cell-mapping-f9pkg" Sep 30 14:16:17 crc kubenswrapper[4840]: I0930 14:16:17.358848 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/521c6360-ed9d-4b6d-a74b-188e3a27be00-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-f9pkg\" (UID: \"521c6360-ed9d-4b6d-a74b-188e3a27be00\") " pod="openstack/nova-cell1-cell-mapping-f9pkg" Sep 30 14:16:17 crc kubenswrapper[4840]: I0930 14:16:17.358991 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpzhn\" (UniqueName: \"kubernetes.io/projected/521c6360-ed9d-4b6d-a74b-188e3a27be00-kube-api-access-fpzhn\") pod \"nova-cell1-cell-mapping-f9pkg\" (UID: \"521c6360-ed9d-4b6d-a74b-188e3a27be00\") " pod="openstack/nova-cell1-cell-mapping-f9pkg" Sep 30 14:16:17 crc kubenswrapper[4840]: I0930 14:16:17.359065 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/521c6360-ed9d-4b6d-a74b-188e3a27be00-config-data\") pod \"nova-cell1-cell-mapping-f9pkg\" (UID: \"521c6360-ed9d-4b6d-a74b-188e3a27be00\") " pod="openstack/nova-cell1-cell-mapping-f9pkg" Sep 30 14:16:17 crc kubenswrapper[4840]: I0930 14:16:17.460899 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/521c6360-ed9d-4b6d-a74b-188e3a27be00-config-data\") pod \"nova-cell1-cell-mapping-f9pkg\" (UID: \"521c6360-ed9d-4b6d-a74b-188e3a27be00\") " pod="openstack/nova-cell1-cell-mapping-f9pkg" Sep 30 14:16:17 crc kubenswrapper[4840]: I0930 14:16:17.461007 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/521c6360-ed9d-4b6d-a74b-188e3a27be00-scripts\") pod \"nova-cell1-cell-mapping-f9pkg\" (UID: \"521c6360-ed9d-4b6d-a74b-188e3a27be00\") " pod="openstack/nova-cell1-cell-mapping-f9pkg" Sep 30 14:16:17 crc kubenswrapper[4840]: I0930 14:16:17.461046 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/521c6360-ed9d-4b6d-a74b-188e3a27be00-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-f9pkg\" (UID: \"521c6360-ed9d-4b6d-a74b-188e3a27be00\") " pod="openstack/nova-cell1-cell-mapping-f9pkg" Sep 30 14:16:17 crc kubenswrapper[4840]: I0930 14:16:17.461161 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpzhn\" (UniqueName: \"kubernetes.io/projected/521c6360-ed9d-4b6d-a74b-188e3a27be00-kube-api-access-fpzhn\") pod \"nova-cell1-cell-mapping-f9pkg\" (UID: \"521c6360-ed9d-4b6d-a74b-188e3a27be00\") " pod="openstack/nova-cell1-cell-mapping-f9pkg" Sep 30 14:16:17 crc kubenswrapper[4840]: I0930 14:16:17.468267 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/521c6360-ed9d-4b6d-a74b-188e3a27be00-scripts\") pod \"nova-cell1-cell-mapping-f9pkg\" (UID: \"521c6360-ed9d-4b6d-a74b-188e3a27be00\") " pod="openstack/nova-cell1-cell-mapping-f9pkg" Sep 30 14:16:17 crc kubenswrapper[4840]: I0930 14:16:17.469630 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/521c6360-ed9d-4b6d-a74b-188e3a27be00-config-data\") pod \"nova-cell1-cell-mapping-f9pkg\" (UID: \"521c6360-ed9d-4b6d-a74b-188e3a27be00\") " pod="openstack/nova-cell1-cell-mapping-f9pkg" Sep 30 14:16:17 crc kubenswrapper[4840]: I0930 14:16:17.477248 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/521c6360-ed9d-4b6d-a74b-188e3a27be00-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-f9pkg\" (UID: \"521c6360-ed9d-4b6d-a74b-188e3a27be00\") " pod="openstack/nova-cell1-cell-mapping-f9pkg" Sep 30 14:16:17 crc kubenswrapper[4840]: I0930 14:16:17.478832 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpzhn\" (UniqueName: \"kubernetes.io/projected/521c6360-ed9d-4b6d-a74b-188e3a27be00-kube-api-access-fpzhn\") pod \"nova-cell1-cell-mapping-f9pkg\" (UID: \"521c6360-ed9d-4b6d-a74b-188e3a27be00\") " pod="openstack/nova-cell1-cell-mapping-f9pkg" Sep 30 14:16:17 crc kubenswrapper[4840]: I0930 14:16:17.518218 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-f9pkg" Sep 30 14:16:18 crc kubenswrapper[4840]: I0930 14:16:18.010610 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-f9pkg"] Sep 30 14:16:18 crc kubenswrapper[4840]: W0930 14:16:18.011109 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod521c6360_ed9d_4b6d_a74b_188e3a27be00.slice/crio-1ef8d366fc3e43c3d5b40ff6614daa51a7b7b2aa16b53666f9a42fb4d31588f2 WatchSource:0}: Error finding container 1ef8d366fc3e43c3d5b40ff6614daa51a7b7b2aa16b53666f9a42fb4d31588f2: Status 404 returned error can't find the container with id 1ef8d366fc3e43c3d5b40ff6614daa51a7b7b2aa16b53666f9a42fb4d31588f2 Sep 30 14:16:18 crc kubenswrapper[4840]: I0930 14:16:18.845862 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-f9pkg" event={"ID":"521c6360-ed9d-4b6d-a74b-188e3a27be00","Type":"ContainerStarted","Data":"beb1299b9489e95f9a0952cc2035d97810aeeae6821fb33d18934840d4a0b784"} Sep 30 14:16:18 crc kubenswrapper[4840]: I0930 14:16:18.845905 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-f9pkg" event={"ID":"521c6360-ed9d-4b6d-a74b-188e3a27be00","Type":"ContainerStarted","Data":"1ef8d366fc3e43c3d5b40ff6614daa51a7b7b2aa16b53666f9a42fb4d31588f2"} Sep 30 14:16:18 crc kubenswrapper[4840]: I0930 14:16:18.900367 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-f9pkg" podStartSLOduration=1.900345348 podStartE2EDuration="1.900345348s" podCreationTimestamp="2025-09-30 14:16:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:16:18.895093949 +0000 UTC m=+1207.524180392" watchObservedRunningTime="2025-09-30 14:16:18.900345348 +0000 UTC m=+1207.529431771" Sep 30 14:16:21 crc kubenswrapper[4840]: I0930 14:16:21.871526 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:16:21 crc kubenswrapper[4840]: I0930 14:16:21.872067 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:16:21 crc kubenswrapper[4840]: I0930 14:16:21.986334 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Sep 30 14:16:23 crc kubenswrapper[4840]: I0930 14:16:23.888964 4840 generic.go:334] "Generic (PLEG): container finished" podID="521c6360-ed9d-4b6d-a74b-188e3a27be00" containerID="beb1299b9489e95f9a0952cc2035d97810aeeae6821fb33d18934840d4a0b784" exitCode=0 Sep 30 14:16:23 crc kubenswrapper[4840]: I0930 14:16:23.889073 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-f9pkg" event={"ID":"521c6360-ed9d-4b6d-a74b-188e3a27be00","Type":"ContainerDied","Data":"beb1299b9489e95f9a0952cc2035d97810aeeae6821fb33d18934840d4a0b784"} Sep 30 14:16:25 crc kubenswrapper[4840]: I0930 14:16:25.215612 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-f9pkg" Sep 30 14:16:25 crc kubenswrapper[4840]: I0930 14:16:25.304680 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/521c6360-ed9d-4b6d-a74b-188e3a27be00-config-data\") pod \"521c6360-ed9d-4b6d-a74b-188e3a27be00\" (UID: \"521c6360-ed9d-4b6d-a74b-188e3a27be00\") " Sep 30 14:16:25 crc kubenswrapper[4840]: I0930 14:16:25.304773 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/521c6360-ed9d-4b6d-a74b-188e3a27be00-scripts\") pod \"521c6360-ed9d-4b6d-a74b-188e3a27be00\" (UID: \"521c6360-ed9d-4b6d-a74b-188e3a27be00\") " Sep 30 14:16:25 crc kubenswrapper[4840]: I0930 14:16:25.304902 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/521c6360-ed9d-4b6d-a74b-188e3a27be00-combined-ca-bundle\") pod \"521c6360-ed9d-4b6d-a74b-188e3a27be00\" (UID: \"521c6360-ed9d-4b6d-a74b-188e3a27be00\") " Sep 30 14:16:25 crc kubenswrapper[4840]: I0930 14:16:25.305029 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpzhn\" (UniqueName: \"kubernetes.io/projected/521c6360-ed9d-4b6d-a74b-188e3a27be00-kube-api-access-fpzhn\") pod \"521c6360-ed9d-4b6d-a74b-188e3a27be00\" (UID: \"521c6360-ed9d-4b6d-a74b-188e3a27be00\") " Sep 30 14:16:25 crc kubenswrapper[4840]: I0930 14:16:25.313967 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/521c6360-ed9d-4b6d-a74b-188e3a27be00-scripts" (OuterVolumeSpecName: "scripts") pod "521c6360-ed9d-4b6d-a74b-188e3a27be00" (UID: "521c6360-ed9d-4b6d-a74b-188e3a27be00"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:16:25 crc kubenswrapper[4840]: I0930 14:16:25.315499 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/521c6360-ed9d-4b6d-a74b-188e3a27be00-kube-api-access-fpzhn" (OuterVolumeSpecName: "kube-api-access-fpzhn") pod "521c6360-ed9d-4b6d-a74b-188e3a27be00" (UID: "521c6360-ed9d-4b6d-a74b-188e3a27be00"). InnerVolumeSpecName "kube-api-access-fpzhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:16:25 crc kubenswrapper[4840]: I0930 14:16:25.343542 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/521c6360-ed9d-4b6d-a74b-188e3a27be00-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "521c6360-ed9d-4b6d-a74b-188e3a27be00" (UID: "521c6360-ed9d-4b6d-a74b-188e3a27be00"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:16:25 crc kubenswrapper[4840]: I0930 14:16:25.344737 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/521c6360-ed9d-4b6d-a74b-188e3a27be00-config-data" (OuterVolumeSpecName: "config-data") pod "521c6360-ed9d-4b6d-a74b-188e3a27be00" (UID: "521c6360-ed9d-4b6d-a74b-188e3a27be00"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:16:25 crc kubenswrapper[4840]: I0930 14:16:25.407392 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpzhn\" (UniqueName: \"kubernetes.io/projected/521c6360-ed9d-4b6d-a74b-188e3a27be00-kube-api-access-fpzhn\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:25 crc kubenswrapper[4840]: I0930 14:16:25.407435 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/521c6360-ed9d-4b6d-a74b-188e3a27be00-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:25 crc kubenswrapper[4840]: I0930 14:16:25.407445 4840 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/521c6360-ed9d-4b6d-a74b-188e3a27be00-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:25 crc kubenswrapper[4840]: I0930 14:16:25.407454 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/521c6360-ed9d-4b6d-a74b-188e3a27be00-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:25 crc kubenswrapper[4840]: I0930 14:16:25.911497 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-f9pkg" event={"ID":"521c6360-ed9d-4b6d-a74b-188e3a27be00","Type":"ContainerDied","Data":"1ef8d366fc3e43c3d5b40ff6614daa51a7b7b2aa16b53666f9a42fb4d31588f2"} Sep 30 14:16:25 crc kubenswrapper[4840]: I0930 14:16:25.911585 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ef8d366fc3e43c3d5b40ff6614daa51a7b7b2aa16b53666f9a42fb4d31588f2" Sep 30 14:16:25 crc kubenswrapper[4840]: I0930 14:16:25.911643 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-f9pkg" Sep 30 14:16:26 crc kubenswrapper[4840]: I0930 14:16:26.081959 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 30 14:16:26 crc kubenswrapper[4840]: I0930 14:16:26.082213 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d816b4be-b475-4a6d-b7c7-e955741fcbba" containerName="nova-api-log" containerID="cri-o://6a13fa59195a1222469f537a473de6efef9f2bfdeb3434d22401e1c2abc9e2b6" gracePeriod=30 Sep 30 14:16:26 crc kubenswrapper[4840]: I0930 14:16:26.082351 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d816b4be-b475-4a6d-b7c7-e955741fcbba" containerName="nova-api-api" containerID="cri-o://1f59be2c7ab21244947c4b598d95af9dff7a71a384e668a7d6d375e330d18043" gracePeriod=30 Sep 30 14:16:26 crc kubenswrapper[4840]: I0930 14:16:26.146939 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 14:16:26 crc kubenswrapper[4840]: I0930 14:16:26.147461 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="1658ae4f-b344-4f58-9b37-ba03da92230f" containerName="nova-scheduler-scheduler" containerID="cri-o://4b5bd4ba815125f172e4e48d936f0189b8e3ee03ecdc094f46ecfad486aaa61b" gracePeriod=30 Sep 30 14:16:26 crc kubenswrapper[4840]: I0930 14:16:26.160057 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 14:16:26 crc kubenswrapper[4840]: I0930 14:16:26.160319 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a14c8d91-8aff-4a17-8c9f-ce3839b3dbed" containerName="nova-metadata-log" containerID="cri-o://f460bec6a12f70ed6c569f9cac301dc1f74fe1dd0ef1a5be64f6c86939b7918b" gracePeriod=30 Sep 30 14:16:26 crc kubenswrapper[4840]: I0930 14:16:26.160720 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a14c8d91-8aff-4a17-8c9f-ce3839b3dbed" containerName="nova-metadata-metadata" containerID="cri-o://acece978aba77ed13654be82a29a690286db631f6fa052a2922f109a7b51e675" gracePeriod=30 Sep 30 14:16:26 crc kubenswrapper[4840]: I0930 14:16:26.921999 4840 generic.go:334] "Generic (PLEG): container finished" podID="d816b4be-b475-4a6d-b7c7-e955741fcbba" containerID="6a13fa59195a1222469f537a473de6efef9f2bfdeb3434d22401e1c2abc9e2b6" exitCode=143 Sep 30 14:16:26 crc kubenswrapper[4840]: I0930 14:16:26.922099 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d816b4be-b475-4a6d-b7c7-e955741fcbba","Type":"ContainerDied","Data":"6a13fa59195a1222469f537a473de6efef9f2bfdeb3434d22401e1c2abc9e2b6"} Sep 30 14:16:26 crc kubenswrapper[4840]: I0930 14:16:26.924378 4840 generic.go:334] "Generic (PLEG): container finished" podID="a14c8d91-8aff-4a17-8c9f-ce3839b3dbed" containerID="f460bec6a12f70ed6c569f9cac301dc1f74fe1dd0ef1a5be64f6c86939b7918b" exitCode=143 Sep 30 14:16:26 crc kubenswrapper[4840]: I0930 14:16:26.924413 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a14c8d91-8aff-4a17-8c9f-ce3839b3dbed","Type":"ContainerDied","Data":"f460bec6a12f70ed6c569f9cac301dc1f74fe1dd0ef1a5be64f6c86939b7918b"} Sep 30 14:16:27 crc kubenswrapper[4840]: E0930 14:16:27.583717 4840 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4b5bd4ba815125f172e4e48d936f0189b8e3ee03ecdc094f46ecfad486aaa61b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Sep 30 14:16:27 crc kubenswrapper[4840]: E0930 14:16:27.585778 4840 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4b5bd4ba815125f172e4e48d936f0189b8e3ee03ecdc094f46ecfad486aaa61b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Sep 30 14:16:27 crc kubenswrapper[4840]: E0930 14:16:27.588065 4840 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4b5bd4ba815125f172e4e48d936f0189b8e3ee03ecdc094f46ecfad486aaa61b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Sep 30 14:16:27 crc kubenswrapper[4840]: E0930 14:16:27.588214 4840 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="1658ae4f-b344-4f58-9b37-ba03da92230f" containerName="nova-scheduler-scheduler" Sep 30 14:16:29 crc kubenswrapper[4840]: I0930 14:16:29.317035 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="a14c8d91-8aff-4a17-8c9f-ce3839b3dbed" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.195:8775/\": read tcp 10.217.0.2:60732->10.217.0.195:8775: read: connection reset by peer" Sep 30 14:16:29 crc kubenswrapper[4840]: I0930 14:16:29.317769 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="a14c8d91-8aff-4a17-8c9f-ce3839b3dbed" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.195:8775/\": read tcp 10.217.0.2:60740->10.217.0.195:8775: read: connection reset by peer" Sep 30 14:16:29 crc kubenswrapper[4840]: I0930 14:16:29.674825 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 14:16:29 crc kubenswrapper[4840]: I0930 14:16:29.787284 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a14c8d91-8aff-4a17-8c9f-ce3839b3dbed-combined-ca-bundle\") pod \"a14c8d91-8aff-4a17-8c9f-ce3839b3dbed\" (UID: \"a14c8d91-8aff-4a17-8c9f-ce3839b3dbed\") " Sep 30 14:16:29 crc kubenswrapper[4840]: I0930 14:16:29.787388 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzf6n\" (UniqueName: \"kubernetes.io/projected/a14c8d91-8aff-4a17-8c9f-ce3839b3dbed-kube-api-access-jzf6n\") pod \"a14c8d91-8aff-4a17-8c9f-ce3839b3dbed\" (UID: \"a14c8d91-8aff-4a17-8c9f-ce3839b3dbed\") " Sep 30 14:16:29 crc kubenswrapper[4840]: I0930 14:16:29.788773 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a14c8d91-8aff-4a17-8c9f-ce3839b3dbed-logs\") pod \"a14c8d91-8aff-4a17-8c9f-ce3839b3dbed\" (UID: \"a14c8d91-8aff-4a17-8c9f-ce3839b3dbed\") " Sep 30 14:16:29 crc kubenswrapper[4840]: I0930 14:16:29.788819 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a14c8d91-8aff-4a17-8c9f-ce3839b3dbed-nova-metadata-tls-certs\") pod \"a14c8d91-8aff-4a17-8c9f-ce3839b3dbed\" (UID: \"a14c8d91-8aff-4a17-8c9f-ce3839b3dbed\") " Sep 30 14:16:29 crc kubenswrapper[4840]: I0930 14:16:29.789017 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a14c8d91-8aff-4a17-8c9f-ce3839b3dbed-config-data\") pod \"a14c8d91-8aff-4a17-8c9f-ce3839b3dbed\" (UID: \"a14c8d91-8aff-4a17-8c9f-ce3839b3dbed\") " Sep 30 14:16:29 crc kubenswrapper[4840]: I0930 14:16:29.790070 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a14c8d91-8aff-4a17-8c9f-ce3839b3dbed-logs" (OuterVolumeSpecName: "logs") pod "a14c8d91-8aff-4a17-8c9f-ce3839b3dbed" (UID: "a14c8d91-8aff-4a17-8c9f-ce3839b3dbed"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:16:29 crc kubenswrapper[4840]: I0930 14:16:29.818750 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a14c8d91-8aff-4a17-8c9f-ce3839b3dbed-kube-api-access-jzf6n" (OuterVolumeSpecName: "kube-api-access-jzf6n") pod "a14c8d91-8aff-4a17-8c9f-ce3839b3dbed" (UID: "a14c8d91-8aff-4a17-8c9f-ce3839b3dbed"). InnerVolumeSpecName "kube-api-access-jzf6n". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:16:29 crc kubenswrapper[4840]: I0930 14:16:29.891783 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzf6n\" (UniqueName: \"kubernetes.io/projected/a14c8d91-8aff-4a17-8c9f-ce3839b3dbed-kube-api-access-jzf6n\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:29 crc kubenswrapper[4840]: I0930 14:16:29.891814 4840 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a14c8d91-8aff-4a17-8c9f-ce3839b3dbed-logs\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:29 crc kubenswrapper[4840]: I0930 14:16:29.897174 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a14c8d91-8aff-4a17-8c9f-ce3839b3dbed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a14c8d91-8aff-4a17-8c9f-ce3839b3dbed" (UID: "a14c8d91-8aff-4a17-8c9f-ce3839b3dbed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:16:29 crc kubenswrapper[4840]: I0930 14:16:29.897250 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a14c8d91-8aff-4a17-8c9f-ce3839b3dbed-config-data" (OuterVolumeSpecName: "config-data") pod "a14c8d91-8aff-4a17-8c9f-ce3839b3dbed" (UID: "a14c8d91-8aff-4a17-8c9f-ce3839b3dbed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:16:29 crc kubenswrapper[4840]: I0930 14:16:29.901781 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a14c8d91-8aff-4a17-8c9f-ce3839b3dbed-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "a14c8d91-8aff-4a17-8c9f-ce3839b3dbed" (UID: "a14c8d91-8aff-4a17-8c9f-ce3839b3dbed"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:16:29 crc kubenswrapper[4840]: I0930 14:16:29.959281 4840 generic.go:334] "Generic (PLEG): container finished" podID="d816b4be-b475-4a6d-b7c7-e955741fcbba" containerID="1f59be2c7ab21244947c4b598d95af9dff7a71a384e668a7d6d375e330d18043" exitCode=0 Sep 30 14:16:29 crc kubenswrapper[4840]: I0930 14:16:29.959395 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d816b4be-b475-4a6d-b7c7-e955741fcbba","Type":"ContainerDied","Data":"1f59be2c7ab21244947c4b598d95af9dff7a71a384e668a7d6d375e330d18043"} Sep 30 14:16:29 crc kubenswrapper[4840]: I0930 14:16:29.959438 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d816b4be-b475-4a6d-b7c7-e955741fcbba","Type":"ContainerDied","Data":"05933c66dad72a17cf2b7c5fa4ac97cf4be31c73690ef1811294e80c7e32499f"} Sep 30 14:16:29 crc kubenswrapper[4840]: I0930 14:16:29.959452 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="05933c66dad72a17cf2b7c5fa4ac97cf4be31c73690ef1811294e80c7e32499f" Sep 30 14:16:29 crc kubenswrapper[4840]: I0930 14:16:29.961854 4840 generic.go:334] "Generic (PLEG): container finished" podID="a14c8d91-8aff-4a17-8c9f-ce3839b3dbed" containerID="acece978aba77ed13654be82a29a690286db631f6fa052a2922f109a7b51e675" exitCode=0 Sep 30 14:16:29 crc kubenswrapper[4840]: I0930 14:16:29.962051 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 14:16:29 crc kubenswrapper[4840]: I0930 14:16:29.961934 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a14c8d91-8aff-4a17-8c9f-ce3839b3dbed","Type":"ContainerDied","Data":"acece978aba77ed13654be82a29a690286db631f6fa052a2922f109a7b51e675"} Sep 30 14:16:29 crc kubenswrapper[4840]: I0930 14:16:29.962973 4840 scope.go:117] "RemoveContainer" containerID="acece978aba77ed13654be82a29a690286db631f6fa052a2922f109a7b51e675" Sep 30 14:16:29 crc kubenswrapper[4840]: I0930 14:16:29.966268 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 14:16:29 crc kubenswrapper[4840]: I0930 14:16:29.962672 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a14c8d91-8aff-4a17-8c9f-ce3839b3dbed","Type":"ContainerDied","Data":"527bb14db4a24724d0ce9906203a5576bd5eea878813d187f574b204c7835ae5"} Sep 30 14:16:29 crc kubenswrapper[4840]: I0930 14:16:29.989153 4840 scope.go:117] "RemoveContainer" containerID="f460bec6a12f70ed6c569f9cac301dc1f74fe1dd0ef1a5be64f6c86939b7918b" Sep 30 14:16:29 crc kubenswrapper[4840]: I0930 14:16:29.992462 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d816b4be-b475-4a6d-b7c7-e955741fcbba-public-tls-certs\") pod \"d816b4be-b475-4a6d-b7c7-e955741fcbba\" (UID: \"d816b4be-b475-4a6d-b7c7-e955741fcbba\") " Sep 30 14:16:29 crc kubenswrapper[4840]: I0930 14:16:29.992517 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d816b4be-b475-4a6d-b7c7-e955741fcbba-combined-ca-bundle\") pod \"d816b4be-b475-4a6d-b7c7-e955741fcbba\" (UID: \"d816b4be-b475-4a6d-b7c7-e955741fcbba\") " Sep 30 14:16:29 crc kubenswrapper[4840]: I0930 14:16:29.992620 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d816b4be-b475-4a6d-b7c7-e955741fcbba-config-data\") pod \"d816b4be-b475-4a6d-b7c7-e955741fcbba\" (UID: \"d816b4be-b475-4a6d-b7c7-e955741fcbba\") " Sep 30 14:16:29 crc kubenswrapper[4840]: I0930 14:16:29.992683 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lsgjn\" (UniqueName: \"kubernetes.io/projected/d816b4be-b475-4a6d-b7c7-e955741fcbba-kube-api-access-lsgjn\") pod \"d816b4be-b475-4a6d-b7c7-e955741fcbba\" (UID: \"d816b4be-b475-4a6d-b7c7-e955741fcbba\") " Sep 30 14:16:29 crc kubenswrapper[4840]: I0930 14:16:29.992714 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d816b4be-b475-4a6d-b7c7-e955741fcbba-logs\") pod \"d816b4be-b475-4a6d-b7c7-e955741fcbba\" (UID: \"d816b4be-b475-4a6d-b7c7-e955741fcbba\") " Sep 30 14:16:29 crc kubenswrapper[4840]: I0930 14:16:29.992745 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d816b4be-b475-4a6d-b7c7-e955741fcbba-internal-tls-certs\") pod \"d816b4be-b475-4a6d-b7c7-e955741fcbba\" (UID: \"d816b4be-b475-4a6d-b7c7-e955741fcbba\") " Sep 30 14:16:29 crc kubenswrapper[4840]: I0930 14:16:29.993179 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a14c8d91-8aff-4a17-8c9f-ce3839b3dbed-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:29 crc kubenswrapper[4840]: I0930 14:16:29.993208 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a14c8d91-8aff-4a17-8c9f-ce3839b3dbed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:29 crc kubenswrapper[4840]: I0930 14:16:29.993223 4840 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a14c8d91-8aff-4a17-8c9f-ce3839b3dbed-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.006474 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d816b4be-b475-4a6d-b7c7-e955741fcbba-logs" (OuterVolumeSpecName: "logs") pod "d816b4be-b475-4a6d-b7c7-e955741fcbba" (UID: "d816b4be-b475-4a6d-b7c7-e955741fcbba"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.009115 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d816b4be-b475-4a6d-b7c7-e955741fcbba-kube-api-access-lsgjn" (OuterVolumeSpecName: "kube-api-access-lsgjn") pod "d816b4be-b475-4a6d-b7c7-e955741fcbba" (UID: "d816b4be-b475-4a6d-b7c7-e955741fcbba"). InnerVolumeSpecName "kube-api-access-lsgjn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.028729 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.030130 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.050482 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d816b4be-b475-4a6d-b7c7-e955741fcbba-config-data" (OuterVolumeSpecName: "config-data") pod "d816b4be-b475-4a6d-b7c7-e955741fcbba" (UID: "d816b4be-b475-4a6d-b7c7-e955741fcbba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.051511 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Sep 30 14:16:30 crc kubenswrapper[4840]: E0930 14:16:30.057938 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d816b4be-b475-4a6d-b7c7-e955741fcbba" containerName="nova-api-api" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.057970 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="d816b4be-b475-4a6d-b7c7-e955741fcbba" containerName="nova-api-api" Sep 30 14:16:30 crc kubenswrapper[4840]: E0930 14:16:30.058001 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d816b4be-b475-4a6d-b7c7-e955741fcbba" containerName="nova-api-log" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.058009 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="d816b4be-b475-4a6d-b7c7-e955741fcbba" containerName="nova-api-log" Sep 30 14:16:30 crc kubenswrapper[4840]: E0930 14:16:30.058031 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a14c8d91-8aff-4a17-8c9f-ce3839b3dbed" containerName="nova-metadata-metadata" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.058038 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="a14c8d91-8aff-4a17-8c9f-ce3839b3dbed" containerName="nova-metadata-metadata" Sep 30 14:16:30 crc kubenswrapper[4840]: E0930 14:16:30.058050 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="521c6360-ed9d-4b6d-a74b-188e3a27be00" containerName="nova-manage" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.058057 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="521c6360-ed9d-4b6d-a74b-188e3a27be00" containerName="nova-manage" Sep 30 14:16:30 crc kubenswrapper[4840]: E0930 14:16:30.058068 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a14c8d91-8aff-4a17-8c9f-ce3839b3dbed" containerName="nova-metadata-log" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.058075 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="a14c8d91-8aff-4a17-8c9f-ce3839b3dbed" containerName="nova-metadata-log" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.058336 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="a14c8d91-8aff-4a17-8c9f-ce3839b3dbed" containerName="nova-metadata-metadata" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.058356 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="d816b4be-b475-4a6d-b7c7-e955741fcbba" containerName="nova-api-api" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.058366 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="521c6360-ed9d-4b6d-a74b-188e3a27be00" containerName="nova-manage" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.058379 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="d816b4be-b475-4a6d-b7c7-e955741fcbba" containerName="nova-api-log" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.058396 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="a14c8d91-8aff-4a17-8c9f-ce3839b3dbed" containerName="nova-metadata-log" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.059746 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.067320 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.070263 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.082001 4840 scope.go:117] "RemoveContainer" containerID="acece978aba77ed13654be82a29a690286db631f6fa052a2922f109a7b51e675" Sep 30 14:16:30 crc kubenswrapper[4840]: E0930 14:16:30.091344 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"acece978aba77ed13654be82a29a690286db631f6fa052a2922f109a7b51e675\": container with ID starting with acece978aba77ed13654be82a29a690286db631f6fa052a2922f109a7b51e675 not found: ID does not exist" containerID="acece978aba77ed13654be82a29a690286db631f6fa052a2922f109a7b51e675" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.091608 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acece978aba77ed13654be82a29a690286db631f6fa052a2922f109a7b51e675"} err="failed to get container status \"acece978aba77ed13654be82a29a690286db631f6fa052a2922f109a7b51e675\": rpc error: code = NotFound desc = could not find container \"acece978aba77ed13654be82a29a690286db631f6fa052a2922f109a7b51e675\": container with ID starting with acece978aba77ed13654be82a29a690286db631f6fa052a2922f109a7b51e675 not found: ID does not exist" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.091651 4840 scope.go:117] "RemoveContainer" containerID="f460bec6a12f70ed6c569f9cac301dc1f74fe1dd0ef1a5be64f6c86939b7918b" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.099916 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f1f889e-0a43-436f-adfe-6bb0a8e9b909-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0f1f889e-0a43-436f-adfe-6bb0a8e9b909\") " pod="openstack/nova-metadata-0" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.104089 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f1f889e-0a43-436f-adfe-6bb0a8e9b909-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0f1f889e-0a43-436f-adfe-6bb0a8e9b909\") " pod="openstack/nova-metadata-0" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.104303 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f1f889e-0a43-436f-adfe-6bb0a8e9b909-logs\") pod \"nova-metadata-0\" (UID: \"0f1f889e-0a43-436f-adfe-6bb0a8e9b909\") " pod="openstack/nova-metadata-0" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.104798 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.104857 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f1f889e-0a43-436f-adfe-6bb0a8e9b909-config-data\") pod \"nova-metadata-0\" (UID: \"0f1f889e-0a43-436f-adfe-6bb0a8e9b909\") " pod="openstack/nova-metadata-0" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.104917 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7b8d\" (UniqueName: \"kubernetes.io/projected/0f1f889e-0a43-436f-adfe-6bb0a8e9b909-kube-api-access-r7b8d\") pod \"nova-metadata-0\" (UID: \"0f1f889e-0a43-436f-adfe-6bb0a8e9b909\") " pod="openstack/nova-metadata-0" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.105030 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d816b4be-b475-4a6d-b7c7-e955741fcbba-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.105048 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lsgjn\" (UniqueName: \"kubernetes.io/projected/d816b4be-b475-4a6d-b7c7-e955741fcbba-kube-api-access-lsgjn\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.105060 4840 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d816b4be-b475-4a6d-b7c7-e955741fcbba-logs\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:30 crc kubenswrapper[4840]: E0930 14:16:30.142130 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f460bec6a12f70ed6c569f9cac301dc1f74fe1dd0ef1a5be64f6c86939b7918b\": container with ID starting with f460bec6a12f70ed6c569f9cac301dc1f74fe1dd0ef1a5be64f6c86939b7918b not found: ID does not exist" containerID="f460bec6a12f70ed6c569f9cac301dc1f74fe1dd0ef1a5be64f6c86939b7918b" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.142267 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f460bec6a12f70ed6c569f9cac301dc1f74fe1dd0ef1a5be64f6c86939b7918b"} err="failed to get container status \"f460bec6a12f70ed6c569f9cac301dc1f74fe1dd0ef1a5be64f6c86939b7918b\": rpc error: code = NotFound desc = could not find container \"f460bec6a12f70ed6c569f9cac301dc1f74fe1dd0ef1a5be64f6c86939b7918b\": container with ID starting with f460bec6a12f70ed6c569f9cac301dc1f74fe1dd0ef1a5be64f6c86939b7918b not found: ID does not exist" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.206823 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d816b4be-b475-4a6d-b7c7-e955741fcbba-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d816b4be-b475-4a6d-b7c7-e955741fcbba" (UID: "d816b4be-b475-4a6d-b7c7-e955741fcbba"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.208451 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f1f889e-0a43-436f-adfe-6bb0a8e9b909-config-data\") pod \"nova-metadata-0\" (UID: \"0f1f889e-0a43-436f-adfe-6bb0a8e9b909\") " pod="openstack/nova-metadata-0" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.208524 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7b8d\" (UniqueName: \"kubernetes.io/projected/0f1f889e-0a43-436f-adfe-6bb0a8e9b909-kube-api-access-r7b8d\") pod \"nova-metadata-0\" (UID: \"0f1f889e-0a43-436f-adfe-6bb0a8e9b909\") " pod="openstack/nova-metadata-0" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.212599 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f1f889e-0a43-436f-adfe-6bb0a8e9b909-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0f1f889e-0a43-436f-adfe-6bb0a8e9b909\") " pod="openstack/nova-metadata-0" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.212969 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f1f889e-0a43-436f-adfe-6bb0a8e9b909-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0f1f889e-0a43-436f-adfe-6bb0a8e9b909\") " pod="openstack/nova-metadata-0" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.227493 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f1f889e-0a43-436f-adfe-6bb0a8e9b909-logs\") pod \"nova-metadata-0\" (UID: \"0f1f889e-0a43-436f-adfe-6bb0a8e9b909\") " pod="openstack/nova-metadata-0" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.228020 4840 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d816b4be-b475-4a6d-b7c7-e955741fcbba-public-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.283396 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f1f889e-0a43-436f-adfe-6bb0a8e9b909-logs\") pod \"nova-metadata-0\" (UID: \"0f1f889e-0a43-436f-adfe-6bb0a8e9b909\") " pod="openstack/nova-metadata-0" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.284316 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a14c8d91-8aff-4a17-8c9f-ce3839b3dbed" path="/var/lib/kubelet/pods/a14c8d91-8aff-4a17-8c9f-ce3839b3dbed/volumes" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.285364 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f1f889e-0a43-436f-adfe-6bb0a8e9b909-config-data\") pod \"nova-metadata-0\" (UID: \"0f1f889e-0a43-436f-adfe-6bb0a8e9b909\") " pod="openstack/nova-metadata-0" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.286245 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f1f889e-0a43-436f-adfe-6bb0a8e9b909-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0f1f889e-0a43-436f-adfe-6bb0a8e9b909\") " pod="openstack/nova-metadata-0" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.288952 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d816b4be-b475-4a6d-b7c7-e955741fcbba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d816b4be-b475-4a6d-b7c7-e955741fcbba" (UID: "d816b4be-b475-4a6d-b7c7-e955741fcbba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.289498 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d816b4be-b475-4a6d-b7c7-e955741fcbba-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d816b4be-b475-4a6d-b7c7-e955741fcbba" (UID: "d816b4be-b475-4a6d-b7c7-e955741fcbba"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.290759 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f1f889e-0a43-436f-adfe-6bb0a8e9b909-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0f1f889e-0a43-436f-adfe-6bb0a8e9b909\") " pod="openstack/nova-metadata-0" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.292376 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7b8d\" (UniqueName: \"kubernetes.io/projected/0f1f889e-0a43-436f-adfe-6bb0a8e9b909-kube-api-access-r7b8d\") pod \"nova-metadata-0\" (UID: \"0f1f889e-0a43-436f-adfe-6bb0a8e9b909\") " pod="openstack/nova-metadata-0" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.332723 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d816b4be-b475-4a6d-b7c7-e955741fcbba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.332756 4840 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d816b4be-b475-4a6d-b7c7-e955741fcbba-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.389654 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.845385 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.972962 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0f1f889e-0a43-436f-adfe-6bb0a8e9b909","Type":"ContainerStarted","Data":"1a02f1f59474b77af663ef9d6a1ae9af4512cc75709a9d41030f8371b2df1a7d"} Sep 30 14:16:30 crc kubenswrapper[4840]: I0930 14:16:30.974810 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 14:16:31 crc kubenswrapper[4840]: I0930 14:16:31.021182 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 30 14:16:31 crc kubenswrapper[4840]: I0930 14:16:31.029253 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Sep 30 14:16:31 crc kubenswrapper[4840]: I0930 14:16:31.044615 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Sep 30 14:16:31 crc kubenswrapper[4840]: I0930 14:16:31.047187 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 14:16:31 crc kubenswrapper[4840]: I0930 14:16:31.055955 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Sep 30 14:16:31 crc kubenswrapper[4840]: I0930 14:16:31.056341 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Sep 30 14:16:31 crc kubenswrapper[4840]: I0930 14:16:31.056350 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Sep 30 14:16:31 crc kubenswrapper[4840]: I0930 14:16:31.086822 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 30 14:16:31 crc kubenswrapper[4840]: I0930 14:16:31.147492 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fl85\" (UniqueName: \"kubernetes.io/projected/eade78f5-391f-455f-b61c-4300f8d9f53c-kube-api-access-4fl85\") pod \"nova-api-0\" (UID: \"eade78f5-391f-455f-b61c-4300f8d9f53c\") " pod="openstack/nova-api-0" Sep 30 14:16:31 crc kubenswrapper[4840]: I0930 14:16:31.147617 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eade78f5-391f-455f-b61c-4300f8d9f53c-logs\") pod \"nova-api-0\" (UID: \"eade78f5-391f-455f-b61c-4300f8d9f53c\") " pod="openstack/nova-api-0" Sep 30 14:16:31 crc kubenswrapper[4840]: I0930 14:16:31.147860 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eade78f5-391f-455f-b61c-4300f8d9f53c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"eade78f5-391f-455f-b61c-4300f8d9f53c\") " pod="openstack/nova-api-0" Sep 30 14:16:31 crc kubenswrapper[4840]: I0930 14:16:31.147918 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eade78f5-391f-455f-b61c-4300f8d9f53c-public-tls-certs\") pod \"nova-api-0\" (UID: \"eade78f5-391f-455f-b61c-4300f8d9f53c\") " pod="openstack/nova-api-0" Sep 30 14:16:31 crc kubenswrapper[4840]: I0930 14:16:31.147966 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eade78f5-391f-455f-b61c-4300f8d9f53c-config-data\") pod \"nova-api-0\" (UID: \"eade78f5-391f-455f-b61c-4300f8d9f53c\") " pod="openstack/nova-api-0" Sep 30 14:16:31 crc kubenswrapper[4840]: I0930 14:16:31.148011 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eade78f5-391f-455f-b61c-4300f8d9f53c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"eade78f5-391f-455f-b61c-4300f8d9f53c\") " pod="openstack/nova-api-0" Sep 30 14:16:31 crc kubenswrapper[4840]: I0930 14:16:31.250258 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eade78f5-391f-455f-b61c-4300f8d9f53c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"eade78f5-391f-455f-b61c-4300f8d9f53c\") " pod="openstack/nova-api-0" Sep 30 14:16:31 crc kubenswrapper[4840]: I0930 14:16:31.250370 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eade78f5-391f-455f-b61c-4300f8d9f53c-public-tls-certs\") pod \"nova-api-0\" (UID: \"eade78f5-391f-455f-b61c-4300f8d9f53c\") " pod="openstack/nova-api-0" Sep 30 14:16:31 crc kubenswrapper[4840]: I0930 14:16:31.250483 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eade78f5-391f-455f-b61c-4300f8d9f53c-config-data\") pod \"nova-api-0\" (UID: \"eade78f5-391f-455f-b61c-4300f8d9f53c\") " pod="openstack/nova-api-0" Sep 30 14:16:31 crc kubenswrapper[4840]: I0930 14:16:31.250693 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eade78f5-391f-455f-b61c-4300f8d9f53c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"eade78f5-391f-455f-b61c-4300f8d9f53c\") " pod="openstack/nova-api-0" Sep 30 14:16:31 crc kubenswrapper[4840]: I0930 14:16:31.250905 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fl85\" (UniqueName: \"kubernetes.io/projected/eade78f5-391f-455f-b61c-4300f8d9f53c-kube-api-access-4fl85\") pod \"nova-api-0\" (UID: \"eade78f5-391f-455f-b61c-4300f8d9f53c\") " pod="openstack/nova-api-0" Sep 30 14:16:31 crc kubenswrapper[4840]: I0930 14:16:31.251318 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eade78f5-391f-455f-b61c-4300f8d9f53c-logs\") pod \"nova-api-0\" (UID: \"eade78f5-391f-455f-b61c-4300f8d9f53c\") " pod="openstack/nova-api-0" Sep 30 14:16:31 crc kubenswrapper[4840]: I0930 14:16:31.253433 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eade78f5-391f-455f-b61c-4300f8d9f53c-logs\") pod \"nova-api-0\" (UID: \"eade78f5-391f-455f-b61c-4300f8d9f53c\") " pod="openstack/nova-api-0" Sep 30 14:16:31 crc kubenswrapper[4840]: I0930 14:16:31.255310 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eade78f5-391f-455f-b61c-4300f8d9f53c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"eade78f5-391f-455f-b61c-4300f8d9f53c\") " pod="openstack/nova-api-0" Sep 30 14:16:31 crc kubenswrapper[4840]: I0930 14:16:31.257494 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eade78f5-391f-455f-b61c-4300f8d9f53c-config-data\") pod \"nova-api-0\" (UID: \"eade78f5-391f-455f-b61c-4300f8d9f53c\") " pod="openstack/nova-api-0" Sep 30 14:16:31 crc kubenswrapper[4840]: I0930 14:16:31.258165 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eade78f5-391f-455f-b61c-4300f8d9f53c-public-tls-certs\") pod \"nova-api-0\" (UID: \"eade78f5-391f-455f-b61c-4300f8d9f53c\") " pod="openstack/nova-api-0" Sep 30 14:16:31 crc kubenswrapper[4840]: I0930 14:16:31.258518 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eade78f5-391f-455f-b61c-4300f8d9f53c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"eade78f5-391f-455f-b61c-4300f8d9f53c\") " pod="openstack/nova-api-0" Sep 30 14:16:31 crc kubenswrapper[4840]: I0930 14:16:31.272158 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fl85\" (UniqueName: \"kubernetes.io/projected/eade78f5-391f-455f-b61c-4300f8d9f53c-kube-api-access-4fl85\") pod \"nova-api-0\" (UID: \"eade78f5-391f-455f-b61c-4300f8d9f53c\") " pod="openstack/nova-api-0" Sep 30 14:16:31 crc kubenswrapper[4840]: I0930 14:16:31.437607 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 14:16:31 crc kubenswrapper[4840]: I0930 14:16:31.912252 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 30 14:16:32 crc kubenswrapper[4840]: I0930 14:16:32.001455 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 30 14:16:32 crc kubenswrapper[4840]: I0930 14:16:32.005148 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"eade78f5-391f-455f-b61c-4300f8d9f53c","Type":"ContainerStarted","Data":"0a1f6ad77020d4bf4838c32409da799b9f2573ef6b5731f32e5ab80d84df6985"} Sep 30 14:16:32 crc kubenswrapper[4840]: I0930 14:16:32.007371 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0f1f889e-0a43-436f-adfe-6bb0a8e9b909","Type":"ContainerStarted","Data":"d153af021c7f14e8f631eda77a359dbda85f59d9fc8b40b4d0a52ee97e68a4f5"} Sep 30 14:16:32 crc kubenswrapper[4840]: I0930 14:16:32.007407 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0f1f889e-0a43-436f-adfe-6bb0a8e9b909","Type":"ContainerStarted","Data":"74a069236b772355326c19e39367f88e8f0fb92d56afae24bf322df162b1be30"} Sep 30 14:16:32 crc kubenswrapper[4840]: I0930 14:16:32.034327 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.034302475 podStartE2EDuration="2.034302475s" podCreationTimestamp="2025-09-30 14:16:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:16:32.025112653 +0000 UTC m=+1220.654199096" watchObservedRunningTime="2025-09-30 14:16:32.034302475 +0000 UTC m=+1220.663388898" Sep 30 14:16:32 crc kubenswrapper[4840]: I0930 14:16:32.131292 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d816b4be-b475-4a6d-b7c7-e955741fcbba" path="/var/lib/kubelet/pods/d816b4be-b475-4a6d-b7c7-e955741fcbba/volumes" Sep 30 14:16:32 crc kubenswrapper[4840]: E0930 14:16:32.583208 4840 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4b5bd4ba815125f172e4e48d936f0189b8e3ee03ecdc094f46ecfad486aaa61b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Sep 30 14:16:32 crc kubenswrapper[4840]: E0930 14:16:32.604374 4840 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4b5bd4ba815125f172e4e48d936f0189b8e3ee03ecdc094f46ecfad486aaa61b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Sep 30 14:16:32 crc kubenswrapper[4840]: E0930 14:16:32.609756 4840 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4b5bd4ba815125f172e4e48d936f0189b8e3ee03ecdc094f46ecfad486aaa61b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Sep 30 14:16:32 crc kubenswrapper[4840]: E0930 14:16:32.609855 4840 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="1658ae4f-b344-4f58-9b37-ba03da92230f" containerName="nova-scheduler-scheduler" Sep 30 14:16:32 crc kubenswrapper[4840]: I0930 14:16:32.965550 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 30 14:16:33 crc kubenswrapper[4840]: I0930 14:16:33.018061 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"eade78f5-391f-455f-b61c-4300f8d9f53c","Type":"ContainerStarted","Data":"16684eb553b7e6f3c369b508dc41c55495df01c8718f6f8dfd7f241d7d896919"} Sep 30 14:16:33 crc kubenswrapper[4840]: I0930 14:16:33.018120 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"eade78f5-391f-455f-b61c-4300f8d9f53c","Type":"ContainerStarted","Data":"f9f245eee24264cf5a05f95744447ef578fccaf825cc550c75a24990c4bb58ba"} Sep 30 14:16:33 crc kubenswrapper[4840]: I0930 14:16:33.045469 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.045447859 podStartE2EDuration="2.045447859s" podCreationTimestamp="2025-09-30 14:16:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:16:33.038930807 +0000 UTC m=+1221.668017240" watchObservedRunningTime="2025-09-30 14:16:33.045447859 +0000 UTC m=+1221.674534282" Sep 30 14:16:35 crc kubenswrapper[4840]: I0930 14:16:35.390811 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 30 14:16:35 crc kubenswrapper[4840]: I0930 14:16:35.391124 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 30 14:16:36 crc kubenswrapper[4840]: I0930 14:16:36.345909 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="b8c84ed0-e20f-445e-9548-4ce979c23476" containerName="rabbitmq" containerID="cri-o://ae16b156af4a276de23012df6e733c3b5d485fee9106ae2949eef5807da82adb" gracePeriod=604796 Sep 30 14:16:36 crc kubenswrapper[4840]: I0930 14:16:36.724399 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 14:16:37 crc kubenswrapper[4840]: I0930 14:16:36.878053 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2b494\" (UniqueName: \"kubernetes.io/projected/1658ae4f-b344-4f58-9b37-ba03da92230f-kube-api-access-2b494\") pod \"1658ae4f-b344-4f58-9b37-ba03da92230f\" (UID: \"1658ae4f-b344-4f58-9b37-ba03da92230f\") " Sep 30 14:16:37 crc kubenswrapper[4840]: I0930 14:16:36.878261 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1658ae4f-b344-4f58-9b37-ba03da92230f-config-data\") pod \"1658ae4f-b344-4f58-9b37-ba03da92230f\" (UID: \"1658ae4f-b344-4f58-9b37-ba03da92230f\") " Sep 30 14:16:37 crc kubenswrapper[4840]: I0930 14:16:36.878283 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1658ae4f-b344-4f58-9b37-ba03da92230f-combined-ca-bundle\") pod \"1658ae4f-b344-4f58-9b37-ba03da92230f\" (UID: \"1658ae4f-b344-4f58-9b37-ba03da92230f\") " Sep 30 14:16:37 crc kubenswrapper[4840]: I0930 14:16:36.885937 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1658ae4f-b344-4f58-9b37-ba03da92230f-kube-api-access-2b494" (OuterVolumeSpecName: "kube-api-access-2b494") pod "1658ae4f-b344-4f58-9b37-ba03da92230f" (UID: "1658ae4f-b344-4f58-9b37-ba03da92230f"). InnerVolumeSpecName "kube-api-access-2b494". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:16:37 crc kubenswrapper[4840]: I0930 14:16:36.932760 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1658ae4f-b344-4f58-9b37-ba03da92230f-config-data" (OuterVolumeSpecName: "config-data") pod "1658ae4f-b344-4f58-9b37-ba03da92230f" (UID: "1658ae4f-b344-4f58-9b37-ba03da92230f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:16:37 crc kubenswrapper[4840]: I0930 14:16:36.949123 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1658ae4f-b344-4f58-9b37-ba03da92230f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1658ae4f-b344-4f58-9b37-ba03da92230f" (UID: "1658ae4f-b344-4f58-9b37-ba03da92230f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:16:37 crc kubenswrapper[4840]: I0930 14:16:36.980631 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1658ae4f-b344-4f58-9b37-ba03da92230f-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:37 crc kubenswrapper[4840]: I0930 14:16:36.980664 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1658ae4f-b344-4f58-9b37-ba03da92230f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:37 crc kubenswrapper[4840]: I0930 14:16:36.980679 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2b494\" (UniqueName: \"kubernetes.io/projected/1658ae4f-b344-4f58-9b37-ba03da92230f-kube-api-access-2b494\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:37 crc kubenswrapper[4840]: I0930 14:16:37.054404 4840 generic.go:334] "Generic (PLEG): container finished" podID="1658ae4f-b344-4f58-9b37-ba03da92230f" containerID="4b5bd4ba815125f172e4e48d936f0189b8e3ee03ecdc094f46ecfad486aaa61b" exitCode=0 Sep 30 14:16:37 crc kubenswrapper[4840]: I0930 14:16:37.054489 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 14:16:37 crc kubenswrapper[4840]: I0930 14:16:37.054521 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1658ae4f-b344-4f58-9b37-ba03da92230f","Type":"ContainerDied","Data":"4b5bd4ba815125f172e4e48d936f0189b8e3ee03ecdc094f46ecfad486aaa61b"} Sep 30 14:16:37 crc kubenswrapper[4840]: I0930 14:16:37.054890 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1658ae4f-b344-4f58-9b37-ba03da92230f","Type":"ContainerDied","Data":"d2ea6fa32ccbb46846d1c3682f161d6544c4a5b55c470b58269eac5301d3f6a8"} Sep 30 14:16:37 crc kubenswrapper[4840]: I0930 14:16:37.054920 4840 scope.go:117] "RemoveContainer" containerID="4b5bd4ba815125f172e4e48d936f0189b8e3ee03ecdc094f46ecfad486aaa61b" Sep 30 14:16:37 crc kubenswrapper[4840]: I0930 14:16:37.076642 4840 scope.go:117] "RemoveContainer" containerID="4b5bd4ba815125f172e4e48d936f0189b8e3ee03ecdc094f46ecfad486aaa61b" Sep 30 14:16:37 crc kubenswrapper[4840]: E0930 14:16:37.077214 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b5bd4ba815125f172e4e48d936f0189b8e3ee03ecdc094f46ecfad486aaa61b\": container with ID starting with 4b5bd4ba815125f172e4e48d936f0189b8e3ee03ecdc094f46ecfad486aaa61b not found: ID does not exist" containerID="4b5bd4ba815125f172e4e48d936f0189b8e3ee03ecdc094f46ecfad486aaa61b" Sep 30 14:16:37 crc kubenswrapper[4840]: I0930 14:16:37.077251 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b5bd4ba815125f172e4e48d936f0189b8e3ee03ecdc094f46ecfad486aaa61b"} err="failed to get container status \"4b5bd4ba815125f172e4e48d936f0189b8e3ee03ecdc094f46ecfad486aaa61b\": rpc error: code = NotFound desc = could not find container \"4b5bd4ba815125f172e4e48d936f0189b8e3ee03ecdc094f46ecfad486aaa61b\": container with ID starting with 4b5bd4ba815125f172e4e48d936f0189b8e3ee03ecdc094f46ecfad486aaa61b not found: ID does not exist" Sep 30 14:16:37 crc kubenswrapper[4840]: I0930 14:16:37.094387 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 14:16:37 crc kubenswrapper[4840]: I0930 14:16:37.106112 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 14:16:37 crc kubenswrapper[4840]: I0930 14:16:37.116621 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 14:16:37 crc kubenswrapper[4840]: E0930 14:16:37.117098 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1658ae4f-b344-4f58-9b37-ba03da92230f" containerName="nova-scheduler-scheduler" Sep 30 14:16:37 crc kubenswrapper[4840]: I0930 14:16:37.117115 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="1658ae4f-b344-4f58-9b37-ba03da92230f" containerName="nova-scheduler-scheduler" Sep 30 14:16:37 crc kubenswrapper[4840]: I0930 14:16:37.117365 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="1658ae4f-b344-4f58-9b37-ba03da92230f" containerName="nova-scheduler-scheduler" Sep 30 14:16:37 crc kubenswrapper[4840]: I0930 14:16:37.119288 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 14:16:37 crc kubenswrapper[4840]: I0930 14:16:37.121822 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Sep 30 14:16:37 crc kubenswrapper[4840]: I0930 14:16:37.129533 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 14:16:37 crc kubenswrapper[4840]: I0930 14:16:37.284617 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca7038c5-867f-4546-b846-7a6beae1fa1f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ca7038c5-867f-4546-b846-7a6beae1fa1f\") " pod="openstack/nova-scheduler-0" Sep 30 14:16:37 crc kubenswrapper[4840]: I0930 14:16:37.284727 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwchn\" (UniqueName: \"kubernetes.io/projected/ca7038c5-867f-4546-b846-7a6beae1fa1f-kube-api-access-xwchn\") pod \"nova-scheduler-0\" (UID: \"ca7038c5-867f-4546-b846-7a6beae1fa1f\") " pod="openstack/nova-scheduler-0" Sep 30 14:16:37 crc kubenswrapper[4840]: I0930 14:16:37.284836 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca7038c5-867f-4546-b846-7a6beae1fa1f-config-data\") pod \"nova-scheduler-0\" (UID: \"ca7038c5-867f-4546-b846-7a6beae1fa1f\") " pod="openstack/nova-scheduler-0" Sep 30 14:16:37 crc kubenswrapper[4840]: I0930 14:16:37.387246 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwchn\" (UniqueName: \"kubernetes.io/projected/ca7038c5-867f-4546-b846-7a6beae1fa1f-kube-api-access-xwchn\") pod \"nova-scheduler-0\" (UID: \"ca7038c5-867f-4546-b846-7a6beae1fa1f\") " pod="openstack/nova-scheduler-0" Sep 30 14:16:37 crc kubenswrapper[4840]: I0930 14:16:37.387376 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca7038c5-867f-4546-b846-7a6beae1fa1f-config-data\") pod \"nova-scheduler-0\" (UID: \"ca7038c5-867f-4546-b846-7a6beae1fa1f\") " pod="openstack/nova-scheduler-0" Sep 30 14:16:37 crc kubenswrapper[4840]: I0930 14:16:37.387495 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca7038c5-867f-4546-b846-7a6beae1fa1f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ca7038c5-867f-4546-b846-7a6beae1fa1f\") " pod="openstack/nova-scheduler-0" Sep 30 14:16:37 crc kubenswrapper[4840]: I0930 14:16:37.400357 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca7038c5-867f-4546-b846-7a6beae1fa1f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ca7038c5-867f-4546-b846-7a6beae1fa1f\") " pod="openstack/nova-scheduler-0" Sep 30 14:16:37 crc kubenswrapper[4840]: I0930 14:16:37.400382 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca7038c5-867f-4546-b846-7a6beae1fa1f-config-data\") pod \"nova-scheduler-0\" (UID: \"ca7038c5-867f-4546-b846-7a6beae1fa1f\") " pod="openstack/nova-scheduler-0" Sep 30 14:16:37 crc kubenswrapper[4840]: I0930 14:16:37.414495 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwchn\" (UniqueName: \"kubernetes.io/projected/ca7038c5-867f-4546-b846-7a6beae1fa1f-kube-api-access-xwchn\") pod \"nova-scheduler-0\" (UID: \"ca7038c5-867f-4546-b846-7a6beae1fa1f\") " pod="openstack/nova-scheduler-0" Sep 30 14:16:37 crc kubenswrapper[4840]: I0930 14:16:37.451975 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 14:16:37 crc kubenswrapper[4840]: I0930 14:16:37.483004 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="c15070b3-f247-4879-a7dc-618faf7e6e35" containerName="rabbitmq" containerID="cri-o://d4a8b19c278d5dd784385c0c8a86fb6c872488a691d678942412c006d4dd7154" gracePeriod=604796 Sep 30 14:16:37 crc kubenswrapper[4840]: I0930 14:16:37.908885 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 14:16:37 crc kubenswrapper[4840]: W0930 14:16:37.911499 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca7038c5_867f_4546_b846_7a6beae1fa1f.slice/crio-7a66023c366f6efa6a22d31987117d421cb0fd13db418d96627f37461da7026b WatchSource:0}: Error finding container 7a66023c366f6efa6a22d31987117d421cb0fd13db418d96627f37461da7026b: Status 404 returned error can't find the container with id 7a66023c366f6efa6a22d31987117d421cb0fd13db418d96627f37461da7026b Sep 30 14:16:38 crc kubenswrapper[4840]: I0930 14:16:38.070684 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ca7038c5-867f-4546-b846-7a6beae1fa1f","Type":"ContainerStarted","Data":"7a66023c366f6efa6a22d31987117d421cb0fd13db418d96627f37461da7026b"} Sep 30 14:16:38 crc kubenswrapper[4840]: I0930 14:16:38.136886 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1658ae4f-b344-4f58-9b37-ba03da92230f" path="/var/lib/kubelet/pods/1658ae4f-b344-4f58-9b37-ba03da92230f/volumes" Sep 30 14:16:39 crc kubenswrapper[4840]: I0930 14:16:39.081920 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ca7038c5-867f-4546-b846-7a6beae1fa1f","Type":"ContainerStarted","Data":"b97168b499f9ad9b2ed3ba90e8bb3721804e7c49ede42d1bb2360b8fa1eaba86"} Sep 30 14:16:40 crc kubenswrapper[4840]: I0930 14:16:40.390284 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Sep 30 14:16:40 crc kubenswrapper[4840]: I0930 14:16:40.390643 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Sep 30 14:16:40 crc kubenswrapper[4840]: I0930 14:16:40.653917 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="c15070b3-f247-4879-a7dc-618faf7e6e35" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.100:5671: connect: connection refused" Sep 30 14:16:40 crc kubenswrapper[4840]: I0930 14:16:40.900024 4840 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="b8c84ed0-e20f-445e-9548-4ce979c23476" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.101:5671: connect: connection refused" Sep 30 14:16:41 crc kubenswrapper[4840]: I0930 14:16:41.404852 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="0f1f889e-0a43-436f-adfe-6bb0a8e9b909" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 30 14:16:41 crc kubenswrapper[4840]: I0930 14:16:41.404972 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="0f1f889e-0a43-436f-adfe-6bb0a8e9b909" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 30 14:16:41 crc kubenswrapper[4840]: I0930 14:16:41.437967 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 30 14:16:41 crc kubenswrapper[4840]: I0930 14:16:41.438043 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 30 14:16:42 crc kubenswrapper[4840]: I0930 14:16:42.451811 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="eade78f5-391f-455f-b61c-4300f8d9f53c" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.206:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 30 14:16:42 crc kubenswrapper[4840]: I0930 14:16:42.451923 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="eade78f5-391f-455f-b61c-4300f8d9f53c" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.206:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 30 14:16:42 crc kubenswrapper[4840]: I0930 14:16:42.452719 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Sep 30 14:16:42 crc kubenswrapper[4840]: I0930 14:16:42.941005 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 30 14:16:42 crc kubenswrapper[4840]: I0930 14:16:42.959870 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=5.959845068 podStartE2EDuration="5.959845068s" podCreationTimestamp="2025-09-30 14:16:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:16:39.102157682 +0000 UTC m=+1227.731244105" watchObservedRunningTime="2025-09-30 14:16:42.959845068 +0000 UTC m=+1231.588931491" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.108456 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b8c84ed0-e20f-445e-9548-4ce979c23476-rabbitmq-erlang-cookie\") pod \"b8c84ed0-e20f-445e-9548-4ce979c23476\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.108493 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b8c84ed0-e20f-445e-9548-4ce979c23476-pod-info\") pod \"b8c84ed0-e20f-445e-9548-4ce979c23476\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.108524 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b8c84ed0-e20f-445e-9548-4ce979c23476-rabbitmq-plugins\") pod \"b8c84ed0-e20f-445e-9548-4ce979c23476\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.108608 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"b8c84ed0-e20f-445e-9548-4ce979c23476\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.108673 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b8c84ed0-e20f-445e-9548-4ce979c23476-config-data\") pod \"b8c84ed0-e20f-445e-9548-4ce979c23476\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.108725 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b8c84ed0-e20f-445e-9548-4ce979c23476-server-conf\") pod \"b8c84ed0-e20f-445e-9548-4ce979c23476\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.108802 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b8c84ed0-e20f-445e-9548-4ce979c23476-erlang-cookie-secret\") pod \"b8c84ed0-e20f-445e-9548-4ce979c23476\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.108869 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b8c84ed0-e20f-445e-9548-4ce979c23476-plugins-conf\") pod \"b8c84ed0-e20f-445e-9548-4ce979c23476\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.108896 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5pt89\" (UniqueName: \"kubernetes.io/projected/b8c84ed0-e20f-445e-9548-4ce979c23476-kube-api-access-5pt89\") pod \"b8c84ed0-e20f-445e-9548-4ce979c23476\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.108921 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b8c84ed0-e20f-445e-9548-4ce979c23476-rabbitmq-tls\") pod \"b8c84ed0-e20f-445e-9548-4ce979c23476\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.108929 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8c84ed0-e20f-445e-9548-4ce979c23476-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "b8c84ed0-e20f-445e-9548-4ce979c23476" (UID: "b8c84ed0-e20f-445e-9548-4ce979c23476"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.108955 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b8c84ed0-e20f-445e-9548-4ce979c23476-rabbitmq-confd\") pod \"b8c84ed0-e20f-445e-9548-4ce979c23476\" (UID: \"b8c84ed0-e20f-445e-9548-4ce979c23476\") " Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.109777 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8c84ed0-e20f-445e-9548-4ce979c23476-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "b8c84ed0-e20f-445e-9548-4ce979c23476" (UID: "b8c84ed0-e20f-445e-9548-4ce979c23476"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.110033 4840 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b8c84ed0-e20f-445e-9548-4ce979c23476-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.110100 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8c84ed0-e20f-445e-9548-4ce979c23476-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "b8c84ed0-e20f-445e-9548-4ce979c23476" (UID: "b8c84ed0-e20f-445e-9548-4ce979c23476"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.115359 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/b8c84ed0-e20f-445e-9548-4ce979c23476-pod-info" (OuterVolumeSpecName: "pod-info") pod "b8c84ed0-e20f-445e-9548-4ce979c23476" (UID: "b8c84ed0-e20f-445e-9548-4ce979c23476"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.131807 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8c84ed0-e20f-445e-9548-4ce979c23476-kube-api-access-5pt89" (OuterVolumeSpecName: "kube-api-access-5pt89") pod "b8c84ed0-e20f-445e-9548-4ce979c23476" (UID: "b8c84ed0-e20f-445e-9548-4ce979c23476"). InnerVolumeSpecName "kube-api-access-5pt89". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.132898 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "b8c84ed0-e20f-445e-9548-4ce979c23476" (UID: "b8c84ed0-e20f-445e-9548-4ce979c23476"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.132950 4840 generic.go:334] "Generic (PLEG): container finished" podID="b8c84ed0-e20f-445e-9548-4ce979c23476" containerID="ae16b156af4a276de23012df6e733c3b5d485fee9106ae2949eef5807da82adb" exitCode=0 Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.132992 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b8c84ed0-e20f-445e-9548-4ce979c23476","Type":"ContainerDied","Data":"ae16b156af4a276de23012df6e733c3b5d485fee9106ae2949eef5807da82adb"} Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.133046 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b8c84ed0-e20f-445e-9548-4ce979c23476","Type":"ContainerDied","Data":"f1a8d380d10fbaae1dc0b42b9d1439d3ba3a55ab28aff1f3e64d2fb16cbf1f05"} Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.133071 4840 scope.go:117] "RemoveContainer" containerID="ae16b156af4a276de23012df6e733c3b5d485fee9106ae2949eef5807da82adb" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.133323 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.140512 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8c84ed0-e20f-445e-9548-4ce979c23476-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "b8c84ed0-e20f-445e-9548-4ce979c23476" (UID: "b8c84ed0-e20f-445e-9548-4ce979c23476"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.144755 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8c84ed0-e20f-445e-9548-4ce979c23476-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "b8c84ed0-e20f-445e-9548-4ce979c23476" (UID: "b8c84ed0-e20f-445e-9548-4ce979c23476"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.170050 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8c84ed0-e20f-445e-9548-4ce979c23476-config-data" (OuterVolumeSpecName: "config-data") pod "b8c84ed0-e20f-445e-9548-4ce979c23476" (UID: "b8c84ed0-e20f-445e-9548-4ce979c23476"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.212779 4840 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b8c84ed0-e20f-445e-9548-4ce979c23476-plugins-conf\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.212805 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5pt89\" (UniqueName: \"kubernetes.io/projected/b8c84ed0-e20f-445e-9548-4ce979c23476-kube-api-access-5pt89\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.212815 4840 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b8c84ed0-e20f-445e-9548-4ce979c23476-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.212824 4840 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b8c84ed0-e20f-445e-9548-4ce979c23476-pod-info\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.212833 4840 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b8c84ed0-e20f-445e-9548-4ce979c23476-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.212855 4840 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.212866 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b8c84ed0-e20f-445e-9548-4ce979c23476-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.212875 4840 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b8c84ed0-e20f-445e-9548-4ce979c23476-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.224078 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8c84ed0-e20f-445e-9548-4ce979c23476-server-conf" (OuterVolumeSpecName: "server-conf") pod "b8c84ed0-e20f-445e-9548-4ce979c23476" (UID: "b8c84ed0-e20f-445e-9548-4ce979c23476"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.246878 4840 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.274799 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8c84ed0-e20f-445e-9548-4ce979c23476-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "b8c84ed0-e20f-445e-9548-4ce979c23476" (UID: "b8c84ed0-e20f-445e-9548-4ce979c23476"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.279685 4840 scope.go:117] "RemoveContainer" containerID="d2cdd432cff6329b6a00d90cea6d034e25aced440092be6ed044b88f6112f277" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.303930 4840 scope.go:117] "RemoveContainer" containerID="ae16b156af4a276de23012df6e733c3b5d485fee9106ae2949eef5807da82adb" Sep 30 14:16:43 crc kubenswrapper[4840]: E0930 14:16:43.304482 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae16b156af4a276de23012df6e733c3b5d485fee9106ae2949eef5807da82adb\": container with ID starting with ae16b156af4a276de23012df6e733c3b5d485fee9106ae2949eef5807da82adb not found: ID does not exist" containerID="ae16b156af4a276de23012df6e733c3b5d485fee9106ae2949eef5807da82adb" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.304533 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae16b156af4a276de23012df6e733c3b5d485fee9106ae2949eef5807da82adb"} err="failed to get container status \"ae16b156af4a276de23012df6e733c3b5d485fee9106ae2949eef5807da82adb\": rpc error: code = NotFound desc = could not find container \"ae16b156af4a276de23012df6e733c3b5d485fee9106ae2949eef5807da82adb\": container with ID starting with ae16b156af4a276de23012df6e733c3b5d485fee9106ae2949eef5807da82adb not found: ID does not exist" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.304713 4840 scope.go:117] "RemoveContainer" containerID="d2cdd432cff6329b6a00d90cea6d034e25aced440092be6ed044b88f6112f277" Sep 30 14:16:43 crc kubenswrapper[4840]: E0930 14:16:43.305113 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2cdd432cff6329b6a00d90cea6d034e25aced440092be6ed044b88f6112f277\": container with ID starting with d2cdd432cff6329b6a00d90cea6d034e25aced440092be6ed044b88f6112f277 not found: ID does not exist" containerID="d2cdd432cff6329b6a00d90cea6d034e25aced440092be6ed044b88f6112f277" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.305163 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2cdd432cff6329b6a00d90cea6d034e25aced440092be6ed044b88f6112f277"} err="failed to get container status \"d2cdd432cff6329b6a00d90cea6d034e25aced440092be6ed044b88f6112f277\": rpc error: code = NotFound desc = could not find container \"d2cdd432cff6329b6a00d90cea6d034e25aced440092be6ed044b88f6112f277\": container with ID starting with d2cdd432cff6329b6a00d90cea6d034e25aced440092be6ed044b88f6112f277 not found: ID does not exist" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.316815 4840 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b8c84ed0-e20f-445e-9548-4ce979c23476-server-conf\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.316846 4840 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b8c84ed0-e20f-445e-9548-4ce979c23476-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.316856 4840 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.523756 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.569525 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.597270 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Sep 30 14:16:43 crc kubenswrapper[4840]: E0930 14:16:43.597798 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8c84ed0-e20f-445e-9548-4ce979c23476" containerName="rabbitmq" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.597819 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8c84ed0-e20f-445e-9548-4ce979c23476" containerName="rabbitmq" Sep 30 14:16:43 crc kubenswrapper[4840]: E0930 14:16:43.597848 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8c84ed0-e20f-445e-9548-4ce979c23476" containerName="setup-container" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.597856 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8c84ed0-e20f-445e-9548-4ce979c23476" containerName="setup-container" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.598103 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8c84ed0-e20f-445e-9548-4ce979c23476" containerName="rabbitmq" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.603294 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.610165 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.610471 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.615606 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.615664 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.615761 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.615835 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.615891 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-s4wjb" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.618562 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.736770 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/97b46089-3117-4a2a-b9db-ada9d0daca8b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"97b46089-3117-4a2a-b9db-ada9d0daca8b\") " pod="openstack/rabbitmq-server-0" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.736829 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/97b46089-3117-4a2a-b9db-ada9d0daca8b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"97b46089-3117-4a2a-b9db-ada9d0daca8b\") " pod="openstack/rabbitmq-server-0" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.736875 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/97b46089-3117-4a2a-b9db-ada9d0daca8b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"97b46089-3117-4a2a-b9db-ada9d0daca8b\") " pod="openstack/rabbitmq-server-0" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.736914 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"97b46089-3117-4a2a-b9db-ada9d0daca8b\") " pod="openstack/rabbitmq-server-0" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.736938 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/97b46089-3117-4a2a-b9db-ada9d0daca8b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"97b46089-3117-4a2a-b9db-ada9d0daca8b\") " pod="openstack/rabbitmq-server-0" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.736962 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/97b46089-3117-4a2a-b9db-ada9d0daca8b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"97b46089-3117-4a2a-b9db-ada9d0daca8b\") " pod="openstack/rabbitmq-server-0" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.737009 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmv8v\" (UniqueName: \"kubernetes.io/projected/97b46089-3117-4a2a-b9db-ada9d0daca8b-kube-api-access-tmv8v\") pod \"rabbitmq-server-0\" (UID: \"97b46089-3117-4a2a-b9db-ada9d0daca8b\") " pod="openstack/rabbitmq-server-0" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.737044 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/97b46089-3117-4a2a-b9db-ada9d0daca8b-config-data\") pod \"rabbitmq-server-0\" (UID: \"97b46089-3117-4a2a-b9db-ada9d0daca8b\") " pod="openstack/rabbitmq-server-0" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.737060 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/97b46089-3117-4a2a-b9db-ada9d0daca8b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"97b46089-3117-4a2a-b9db-ada9d0daca8b\") " pod="openstack/rabbitmq-server-0" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.737081 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/97b46089-3117-4a2a-b9db-ada9d0daca8b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"97b46089-3117-4a2a-b9db-ada9d0daca8b\") " pod="openstack/rabbitmq-server-0" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.737099 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/97b46089-3117-4a2a-b9db-ada9d0daca8b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"97b46089-3117-4a2a-b9db-ada9d0daca8b\") " pod="openstack/rabbitmq-server-0" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.838187 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/97b46089-3117-4a2a-b9db-ada9d0daca8b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"97b46089-3117-4a2a-b9db-ada9d0daca8b\") " pod="openstack/rabbitmq-server-0" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.838257 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"97b46089-3117-4a2a-b9db-ada9d0daca8b\") " pod="openstack/rabbitmq-server-0" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.838296 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/97b46089-3117-4a2a-b9db-ada9d0daca8b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"97b46089-3117-4a2a-b9db-ada9d0daca8b\") " pod="openstack/rabbitmq-server-0" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.838327 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/97b46089-3117-4a2a-b9db-ada9d0daca8b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"97b46089-3117-4a2a-b9db-ada9d0daca8b\") " pod="openstack/rabbitmq-server-0" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.838398 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmv8v\" (UniqueName: \"kubernetes.io/projected/97b46089-3117-4a2a-b9db-ada9d0daca8b-kube-api-access-tmv8v\") pod \"rabbitmq-server-0\" (UID: \"97b46089-3117-4a2a-b9db-ada9d0daca8b\") " pod="openstack/rabbitmq-server-0" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.838451 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/97b46089-3117-4a2a-b9db-ada9d0daca8b-config-data\") pod \"rabbitmq-server-0\" (UID: \"97b46089-3117-4a2a-b9db-ada9d0daca8b\") " pod="openstack/rabbitmq-server-0" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.838474 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/97b46089-3117-4a2a-b9db-ada9d0daca8b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"97b46089-3117-4a2a-b9db-ada9d0daca8b\") " pod="openstack/rabbitmq-server-0" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.838503 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/97b46089-3117-4a2a-b9db-ada9d0daca8b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"97b46089-3117-4a2a-b9db-ada9d0daca8b\") " pod="openstack/rabbitmq-server-0" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.838526 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/97b46089-3117-4a2a-b9db-ada9d0daca8b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"97b46089-3117-4a2a-b9db-ada9d0daca8b\") " pod="openstack/rabbitmq-server-0" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.838659 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/97b46089-3117-4a2a-b9db-ada9d0daca8b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"97b46089-3117-4a2a-b9db-ada9d0daca8b\") " pod="openstack/rabbitmq-server-0" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.838705 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/97b46089-3117-4a2a-b9db-ada9d0daca8b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"97b46089-3117-4a2a-b9db-ada9d0daca8b\") " pod="openstack/rabbitmq-server-0" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.839187 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/97b46089-3117-4a2a-b9db-ada9d0daca8b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"97b46089-3117-4a2a-b9db-ada9d0daca8b\") " pod="openstack/rabbitmq-server-0" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.839473 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/97b46089-3117-4a2a-b9db-ada9d0daca8b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"97b46089-3117-4a2a-b9db-ada9d0daca8b\") " pod="openstack/rabbitmq-server-0" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.840257 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/97b46089-3117-4a2a-b9db-ada9d0daca8b-config-data\") pod \"rabbitmq-server-0\" (UID: \"97b46089-3117-4a2a-b9db-ada9d0daca8b\") " pod="openstack/rabbitmq-server-0" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.840353 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/97b46089-3117-4a2a-b9db-ada9d0daca8b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"97b46089-3117-4a2a-b9db-ada9d0daca8b\") " pod="openstack/rabbitmq-server-0" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.840686 4840 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"97b46089-3117-4a2a-b9db-ada9d0daca8b\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-server-0" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.841823 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/97b46089-3117-4a2a-b9db-ada9d0daca8b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"97b46089-3117-4a2a-b9db-ada9d0daca8b\") " pod="openstack/rabbitmq-server-0" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.847578 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/97b46089-3117-4a2a-b9db-ada9d0daca8b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"97b46089-3117-4a2a-b9db-ada9d0daca8b\") " pod="openstack/rabbitmq-server-0" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.856935 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/97b46089-3117-4a2a-b9db-ada9d0daca8b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"97b46089-3117-4a2a-b9db-ada9d0daca8b\") " pod="openstack/rabbitmq-server-0" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.857317 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/97b46089-3117-4a2a-b9db-ada9d0daca8b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"97b46089-3117-4a2a-b9db-ada9d0daca8b\") " pod="openstack/rabbitmq-server-0" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.859381 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/97b46089-3117-4a2a-b9db-ada9d0daca8b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"97b46089-3117-4a2a-b9db-ada9d0daca8b\") " pod="openstack/rabbitmq-server-0" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.864468 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmv8v\" (UniqueName: \"kubernetes.io/projected/97b46089-3117-4a2a-b9db-ada9d0daca8b-kube-api-access-tmv8v\") pod \"rabbitmq-server-0\" (UID: \"97b46089-3117-4a2a-b9db-ada9d0daca8b\") " pod="openstack/rabbitmq-server-0" Sep 30 14:16:43 crc kubenswrapper[4840]: I0930 14:16:43.889026 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"97b46089-3117-4a2a-b9db-ada9d0daca8b\") " pod="openstack/rabbitmq-server-0" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.008710 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.060112 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.144139 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8c84ed0-e20f-445e-9548-4ce979c23476" path="/var/lib/kubelet/pods/b8c84ed0-e20f-445e-9548-4ce979c23476/volumes" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.144421 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c15070b3-f247-4879-a7dc-618faf7e6e35-rabbitmq-confd\") pod \"c15070b3-f247-4879-a7dc-618faf7e6e35\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.144507 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggjnw\" (UniqueName: \"kubernetes.io/projected/c15070b3-f247-4879-a7dc-618faf7e6e35-kube-api-access-ggjnw\") pod \"c15070b3-f247-4879-a7dc-618faf7e6e35\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.144535 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c15070b3-f247-4879-a7dc-618faf7e6e35-plugins-conf\") pod \"c15070b3-f247-4879-a7dc-618faf7e6e35\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.144596 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c15070b3-f247-4879-a7dc-618faf7e6e35-rabbitmq-plugins\") pod \"c15070b3-f247-4879-a7dc-618faf7e6e35\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.144639 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c15070b3-f247-4879-a7dc-618faf7e6e35-rabbitmq-erlang-cookie\") pod \"c15070b3-f247-4879-a7dc-618faf7e6e35\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.144703 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c15070b3-f247-4879-a7dc-618faf7e6e35-config-data\") pod \"c15070b3-f247-4879-a7dc-618faf7e6e35\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.144736 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c15070b3-f247-4879-a7dc-618faf7e6e35-server-conf\") pod \"c15070b3-f247-4879-a7dc-618faf7e6e35\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.144810 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c15070b3-f247-4879-a7dc-618faf7e6e35-rabbitmq-tls\") pod \"c15070b3-f247-4879-a7dc-618faf7e6e35\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.144874 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c15070b3-f247-4879-a7dc-618faf7e6e35-pod-info\") pod \"c15070b3-f247-4879-a7dc-618faf7e6e35\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.144900 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c15070b3-f247-4879-a7dc-618faf7e6e35-erlang-cookie-secret\") pod \"c15070b3-f247-4879-a7dc-618faf7e6e35\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.144935 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"c15070b3-f247-4879-a7dc-618faf7e6e35\" (UID: \"c15070b3-f247-4879-a7dc-618faf7e6e35\") " Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.146091 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c15070b3-f247-4879-a7dc-618faf7e6e35-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "c15070b3-f247-4879-a7dc-618faf7e6e35" (UID: "c15070b3-f247-4879-a7dc-618faf7e6e35"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.154511 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c15070b3-f247-4879-a7dc-618faf7e6e35-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "c15070b3-f247-4879-a7dc-618faf7e6e35" (UID: "c15070b3-f247-4879-a7dc-618faf7e6e35"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.160103 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c15070b3-f247-4879-a7dc-618faf7e6e35-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "c15070b3-f247-4879-a7dc-618faf7e6e35" (UID: "c15070b3-f247-4879-a7dc-618faf7e6e35"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.160908 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c15070b3-f247-4879-a7dc-618faf7e6e35-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "c15070b3-f247-4879-a7dc-618faf7e6e35" (UID: "c15070b3-f247-4879-a7dc-618faf7e6e35"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.161968 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "persistence") pod "c15070b3-f247-4879-a7dc-618faf7e6e35" (UID: "c15070b3-f247-4879-a7dc-618faf7e6e35"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.163840 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c15070b3-f247-4879-a7dc-618faf7e6e35-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "c15070b3-f247-4879-a7dc-618faf7e6e35" (UID: "c15070b3-f247-4879-a7dc-618faf7e6e35"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.165868 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c15070b3-f247-4879-a7dc-618faf7e6e35-kube-api-access-ggjnw" (OuterVolumeSpecName: "kube-api-access-ggjnw") pod "c15070b3-f247-4879-a7dc-618faf7e6e35" (UID: "c15070b3-f247-4879-a7dc-618faf7e6e35"). InnerVolumeSpecName "kube-api-access-ggjnw". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.170675 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/c15070b3-f247-4879-a7dc-618faf7e6e35-pod-info" (OuterVolumeSpecName: "pod-info") pod "c15070b3-f247-4879-a7dc-618faf7e6e35" (UID: "c15070b3-f247-4879-a7dc-618faf7e6e35"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.188320 4840 generic.go:334] "Generic (PLEG): container finished" podID="c15070b3-f247-4879-a7dc-618faf7e6e35" containerID="d4a8b19c278d5dd784385c0c8a86fb6c872488a691d678942412c006d4dd7154" exitCode=0 Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.188383 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c15070b3-f247-4879-a7dc-618faf7e6e35","Type":"ContainerDied","Data":"d4a8b19c278d5dd784385c0c8a86fb6c872488a691d678942412c006d4dd7154"} Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.188408 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c15070b3-f247-4879-a7dc-618faf7e6e35","Type":"ContainerDied","Data":"c4b33017622436b92ab437672fa03b26855f6322498a8fdc967ada50a21cd6f4"} Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.188448 4840 scope.go:117] "RemoveContainer" containerID="d4a8b19c278d5dd784385c0c8a86fb6c872488a691d678942412c006d4dd7154" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.188614 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.210753 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c15070b3-f247-4879-a7dc-618faf7e6e35-config-data" (OuterVolumeSpecName: "config-data") pod "c15070b3-f247-4879-a7dc-618faf7e6e35" (UID: "c15070b3-f247-4879-a7dc-618faf7e6e35"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.248623 4840 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c15070b3-f247-4879-a7dc-618faf7e6e35-pod-info\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.248648 4840 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c15070b3-f247-4879-a7dc-618faf7e6e35-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.248669 4840 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.248678 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggjnw\" (UniqueName: \"kubernetes.io/projected/c15070b3-f247-4879-a7dc-618faf7e6e35-kube-api-access-ggjnw\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.248690 4840 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c15070b3-f247-4879-a7dc-618faf7e6e35-plugins-conf\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.248710 4840 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c15070b3-f247-4879-a7dc-618faf7e6e35-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.248724 4840 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c15070b3-f247-4879-a7dc-618faf7e6e35-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.248736 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c15070b3-f247-4879-a7dc-618faf7e6e35-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.248747 4840 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c15070b3-f247-4879-a7dc-618faf7e6e35-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.263111 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c15070b3-f247-4879-a7dc-618faf7e6e35-server-conf" (OuterVolumeSpecName: "server-conf") pod "c15070b3-f247-4879-a7dc-618faf7e6e35" (UID: "c15070b3-f247-4879-a7dc-618faf7e6e35"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.267577 4840 scope.go:117] "RemoveContainer" containerID="10b078a915c98d66021e33eaf12b5a5033f297335c7c69779a5f20a19252eb1c" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.282675 4840 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.311220 4840 scope.go:117] "RemoveContainer" containerID="d4a8b19c278d5dd784385c0c8a86fb6c872488a691d678942412c006d4dd7154" Sep 30 14:16:44 crc kubenswrapper[4840]: E0930 14:16:44.313645 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4a8b19c278d5dd784385c0c8a86fb6c872488a691d678942412c006d4dd7154\": container with ID starting with d4a8b19c278d5dd784385c0c8a86fb6c872488a691d678942412c006d4dd7154 not found: ID does not exist" containerID="d4a8b19c278d5dd784385c0c8a86fb6c872488a691d678942412c006d4dd7154" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.313686 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4a8b19c278d5dd784385c0c8a86fb6c872488a691d678942412c006d4dd7154"} err="failed to get container status \"d4a8b19c278d5dd784385c0c8a86fb6c872488a691d678942412c006d4dd7154\": rpc error: code = NotFound desc = could not find container \"d4a8b19c278d5dd784385c0c8a86fb6c872488a691d678942412c006d4dd7154\": container with ID starting with d4a8b19c278d5dd784385c0c8a86fb6c872488a691d678942412c006d4dd7154 not found: ID does not exist" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.313715 4840 scope.go:117] "RemoveContainer" containerID="10b078a915c98d66021e33eaf12b5a5033f297335c7c69779a5f20a19252eb1c" Sep 30 14:16:44 crc kubenswrapper[4840]: E0930 14:16:44.314798 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10b078a915c98d66021e33eaf12b5a5033f297335c7c69779a5f20a19252eb1c\": container with ID starting with 10b078a915c98d66021e33eaf12b5a5033f297335c7c69779a5f20a19252eb1c not found: ID does not exist" containerID="10b078a915c98d66021e33eaf12b5a5033f297335c7c69779a5f20a19252eb1c" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.314841 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10b078a915c98d66021e33eaf12b5a5033f297335c7c69779a5f20a19252eb1c"} err="failed to get container status \"10b078a915c98d66021e33eaf12b5a5033f297335c7c69779a5f20a19252eb1c\": rpc error: code = NotFound desc = could not find container \"10b078a915c98d66021e33eaf12b5a5033f297335c7c69779a5f20a19252eb1c\": container with ID starting with 10b078a915c98d66021e33eaf12b5a5033f297335c7c69779a5f20a19252eb1c not found: ID does not exist" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.346777 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c15070b3-f247-4879-a7dc-618faf7e6e35-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "c15070b3-f247-4879-a7dc-618faf7e6e35" (UID: "c15070b3-f247-4879-a7dc-618faf7e6e35"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.350271 4840 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c15070b3-f247-4879-a7dc-618faf7e6e35-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.350301 4840 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c15070b3-f247-4879-a7dc-618faf7e6e35-server-conf\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.350316 4840 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.526524 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.537601 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.553314 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 30 14:16:44 crc kubenswrapper[4840]: E0930 14:16:44.553756 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c15070b3-f247-4879-a7dc-618faf7e6e35" containerName="setup-container" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.553770 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="c15070b3-f247-4879-a7dc-618faf7e6e35" containerName="setup-container" Sep 30 14:16:44 crc kubenswrapper[4840]: E0930 14:16:44.553785 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c15070b3-f247-4879-a7dc-618faf7e6e35" containerName="rabbitmq" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.553792 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="c15070b3-f247-4879-a7dc-618faf7e6e35" containerName="rabbitmq" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.553966 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="c15070b3-f247-4879-a7dc-618faf7e6e35" containerName="rabbitmq" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.555124 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.560697 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.562133 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.562184 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.562206 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.562223 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.562144 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-w66jp" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.562438 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.572817 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 30 14:16:44 crc kubenswrapper[4840]: W0930 14:16:44.575706 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97b46089_3117_4a2a_b9db_ada9d0daca8b.slice/crio-56b39992f846a12f04f296a5edb2d521e71b07715e86dbd9d9d3658fe698fffd WatchSource:0}: Error finding container 56b39992f846a12f04f296a5edb2d521e71b07715e86dbd9d9d3658fe698fffd: Status 404 returned error can't find the container with id 56b39992f846a12f04f296a5edb2d521e71b07715e86dbd9d9d3658fe698fffd Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.581930 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.659620 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/533b8943-b9ea-4062-9dd2-ed1b29d4f2c6-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"533b8943-b9ea-4062-9dd2-ed1b29d4f2c6\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.659714 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/533b8943-b9ea-4062-9dd2-ed1b29d4f2c6-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"533b8943-b9ea-4062-9dd2-ed1b29d4f2c6\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.659759 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/533b8943-b9ea-4062-9dd2-ed1b29d4f2c6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"533b8943-b9ea-4062-9dd2-ed1b29d4f2c6\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.659816 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/533b8943-b9ea-4062-9dd2-ed1b29d4f2c6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"533b8943-b9ea-4062-9dd2-ed1b29d4f2c6\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.659850 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/533b8943-b9ea-4062-9dd2-ed1b29d4f2c6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"533b8943-b9ea-4062-9dd2-ed1b29d4f2c6\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.659878 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/533b8943-b9ea-4062-9dd2-ed1b29d4f2c6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"533b8943-b9ea-4062-9dd2-ed1b29d4f2c6\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.659907 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/533b8943-b9ea-4062-9dd2-ed1b29d4f2c6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"533b8943-b9ea-4062-9dd2-ed1b29d4f2c6\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.660046 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"533b8943-b9ea-4062-9dd2-ed1b29d4f2c6\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.660104 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs2md\" (UniqueName: \"kubernetes.io/projected/533b8943-b9ea-4062-9dd2-ed1b29d4f2c6-kube-api-access-rs2md\") pod \"rabbitmq-cell1-server-0\" (UID: \"533b8943-b9ea-4062-9dd2-ed1b29d4f2c6\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.660130 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/533b8943-b9ea-4062-9dd2-ed1b29d4f2c6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"533b8943-b9ea-4062-9dd2-ed1b29d4f2c6\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.660198 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/533b8943-b9ea-4062-9dd2-ed1b29d4f2c6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"533b8943-b9ea-4062-9dd2-ed1b29d4f2c6\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.761412 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/533b8943-b9ea-4062-9dd2-ed1b29d4f2c6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"533b8943-b9ea-4062-9dd2-ed1b29d4f2c6\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.761473 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/533b8943-b9ea-4062-9dd2-ed1b29d4f2c6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"533b8943-b9ea-4062-9dd2-ed1b29d4f2c6\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.761494 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/533b8943-b9ea-4062-9dd2-ed1b29d4f2c6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"533b8943-b9ea-4062-9dd2-ed1b29d4f2c6\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.761527 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/533b8943-b9ea-4062-9dd2-ed1b29d4f2c6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"533b8943-b9ea-4062-9dd2-ed1b29d4f2c6\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.761633 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"533b8943-b9ea-4062-9dd2-ed1b29d4f2c6\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.761666 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rs2md\" (UniqueName: \"kubernetes.io/projected/533b8943-b9ea-4062-9dd2-ed1b29d4f2c6-kube-api-access-rs2md\") pod \"rabbitmq-cell1-server-0\" (UID: \"533b8943-b9ea-4062-9dd2-ed1b29d4f2c6\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.761683 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/533b8943-b9ea-4062-9dd2-ed1b29d4f2c6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"533b8943-b9ea-4062-9dd2-ed1b29d4f2c6\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.761723 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/533b8943-b9ea-4062-9dd2-ed1b29d4f2c6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"533b8943-b9ea-4062-9dd2-ed1b29d4f2c6\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.761761 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/533b8943-b9ea-4062-9dd2-ed1b29d4f2c6-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"533b8943-b9ea-4062-9dd2-ed1b29d4f2c6\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.761833 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/533b8943-b9ea-4062-9dd2-ed1b29d4f2c6-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"533b8943-b9ea-4062-9dd2-ed1b29d4f2c6\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.761875 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/533b8943-b9ea-4062-9dd2-ed1b29d4f2c6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"533b8943-b9ea-4062-9dd2-ed1b29d4f2c6\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.762013 4840 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"533b8943-b9ea-4062-9dd2-ed1b29d4f2c6\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.763160 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/533b8943-b9ea-4062-9dd2-ed1b29d4f2c6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"533b8943-b9ea-4062-9dd2-ed1b29d4f2c6\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.763190 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/533b8943-b9ea-4062-9dd2-ed1b29d4f2c6-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"533b8943-b9ea-4062-9dd2-ed1b29d4f2c6\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.763431 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/533b8943-b9ea-4062-9dd2-ed1b29d4f2c6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"533b8943-b9ea-4062-9dd2-ed1b29d4f2c6\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.763569 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/533b8943-b9ea-4062-9dd2-ed1b29d4f2c6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"533b8943-b9ea-4062-9dd2-ed1b29d4f2c6\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.763726 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/533b8943-b9ea-4062-9dd2-ed1b29d4f2c6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"533b8943-b9ea-4062-9dd2-ed1b29d4f2c6\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.766378 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/533b8943-b9ea-4062-9dd2-ed1b29d4f2c6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"533b8943-b9ea-4062-9dd2-ed1b29d4f2c6\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.766806 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/533b8943-b9ea-4062-9dd2-ed1b29d4f2c6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"533b8943-b9ea-4062-9dd2-ed1b29d4f2c6\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.767907 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/533b8943-b9ea-4062-9dd2-ed1b29d4f2c6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"533b8943-b9ea-4062-9dd2-ed1b29d4f2c6\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.772479 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/533b8943-b9ea-4062-9dd2-ed1b29d4f2c6-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"533b8943-b9ea-4062-9dd2-ed1b29d4f2c6\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.781202 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs2md\" (UniqueName: \"kubernetes.io/projected/533b8943-b9ea-4062-9dd2-ed1b29d4f2c6-kube-api-access-rs2md\") pod \"rabbitmq-cell1-server-0\" (UID: \"533b8943-b9ea-4062-9dd2-ed1b29d4f2c6\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.805093 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"533b8943-b9ea-4062-9dd2-ed1b29d4f2c6\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:16:44 crc kubenswrapper[4840]: I0930 14:16:44.910856 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:16:45 crc kubenswrapper[4840]: I0930 14:16:45.201953 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"97b46089-3117-4a2a-b9db-ada9d0daca8b","Type":"ContainerStarted","Data":"dcee92a5252d73c0d68f435241b274fd172504116b62bc26d71ec510c69e33a5"} Sep 30 14:16:45 crc kubenswrapper[4840]: I0930 14:16:45.202617 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"97b46089-3117-4a2a-b9db-ada9d0daca8b","Type":"ContainerStarted","Data":"56b39992f846a12f04f296a5edb2d521e71b07715e86dbd9d9d3658fe698fffd"} Sep 30 14:16:45 crc kubenswrapper[4840]: I0930 14:16:45.397118 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 30 14:16:45 crc kubenswrapper[4840]: W0930 14:16:45.407757 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod533b8943_b9ea_4062_9dd2_ed1b29d4f2c6.slice/crio-685109a70fe1f2d4376acdd65ca2953996c834e09892aa99db30cb113f18c510 WatchSource:0}: Error finding container 685109a70fe1f2d4376acdd65ca2953996c834e09892aa99db30cb113f18c510: Status 404 returned error can't find the container with id 685109a70fe1f2d4376acdd65ca2953996c834e09892aa99db30cb113f18c510 Sep 30 14:16:45 crc kubenswrapper[4840]: I0930 14:16:45.933963 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-6wwml"] Sep 30 14:16:45 crc kubenswrapper[4840]: I0930 14:16:45.937100 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-6wwml" Sep 30 14:16:45 crc kubenswrapper[4840]: I0930 14:16:45.939632 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Sep 30 14:16:45 crc kubenswrapper[4840]: I0930 14:16:45.969038 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-6wwml"] Sep 30 14:16:45 crc kubenswrapper[4840]: I0930 14:16:45.985325 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-6wwml\" (UID: \"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a\") " pod="openstack/dnsmasq-dns-67b789f86c-6wwml" Sep 30 14:16:45 crc kubenswrapper[4840]: I0930 14:16:45.985381 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5z2dd\" (UniqueName: \"kubernetes.io/projected/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-kube-api-access-5z2dd\") pod \"dnsmasq-dns-67b789f86c-6wwml\" (UID: \"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a\") " pod="openstack/dnsmasq-dns-67b789f86c-6wwml" Sep 30 14:16:45 crc kubenswrapper[4840]: I0930 14:16:45.985487 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-dns-svc\") pod \"dnsmasq-dns-67b789f86c-6wwml\" (UID: \"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a\") " pod="openstack/dnsmasq-dns-67b789f86c-6wwml" Sep 30 14:16:45 crc kubenswrapper[4840]: I0930 14:16:45.985510 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-6wwml\" (UID: \"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a\") " pod="openstack/dnsmasq-dns-67b789f86c-6wwml" Sep 30 14:16:45 crc kubenswrapper[4840]: I0930 14:16:45.985532 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-6wwml\" (UID: \"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a\") " pod="openstack/dnsmasq-dns-67b789f86c-6wwml" Sep 30 14:16:45 crc kubenswrapper[4840]: I0930 14:16:45.985603 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-config\") pod \"dnsmasq-dns-67b789f86c-6wwml\" (UID: \"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a\") " pod="openstack/dnsmasq-dns-67b789f86c-6wwml" Sep 30 14:16:45 crc kubenswrapper[4840]: I0930 14:16:45.985631 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-6wwml\" (UID: \"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a\") " pod="openstack/dnsmasq-dns-67b789f86c-6wwml" Sep 30 14:16:46 crc kubenswrapper[4840]: I0930 14:16:46.087324 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-6wwml\" (UID: \"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a\") " pod="openstack/dnsmasq-dns-67b789f86c-6wwml" Sep 30 14:16:46 crc kubenswrapper[4840]: I0930 14:16:46.087427 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-config\") pod \"dnsmasq-dns-67b789f86c-6wwml\" (UID: \"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a\") " pod="openstack/dnsmasq-dns-67b789f86c-6wwml" Sep 30 14:16:46 crc kubenswrapper[4840]: I0930 14:16:46.087458 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-6wwml\" (UID: \"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a\") " pod="openstack/dnsmasq-dns-67b789f86c-6wwml" Sep 30 14:16:46 crc kubenswrapper[4840]: I0930 14:16:46.087545 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-6wwml\" (UID: \"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a\") " pod="openstack/dnsmasq-dns-67b789f86c-6wwml" Sep 30 14:16:46 crc kubenswrapper[4840]: I0930 14:16:46.088525 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-config\") pod \"dnsmasq-dns-67b789f86c-6wwml\" (UID: \"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a\") " pod="openstack/dnsmasq-dns-67b789f86c-6wwml" Sep 30 14:16:46 crc kubenswrapper[4840]: I0930 14:16:46.088663 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-6wwml\" (UID: \"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a\") " pod="openstack/dnsmasq-dns-67b789f86c-6wwml" Sep 30 14:16:46 crc kubenswrapper[4840]: I0930 14:16:46.088743 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5z2dd\" (UniqueName: \"kubernetes.io/projected/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-kube-api-access-5z2dd\") pod \"dnsmasq-dns-67b789f86c-6wwml\" (UID: \"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a\") " pod="openstack/dnsmasq-dns-67b789f86c-6wwml" Sep 30 14:16:46 crc kubenswrapper[4840]: I0930 14:16:46.088776 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-6wwml\" (UID: \"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a\") " pod="openstack/dnsmasq-dns-67b789f86c-6wwml" Sep 30 14:16:46 crc kubenswrapper[4840]: I0930 14:16:46.089215 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-6wwml\" (UID: \"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a\") " pod="openstack/dnsmasq-dns-67b789f86c-6wwml" Sep 30 14:16:46 crc kubenswrapper[4840]: I0930 14:16:46.089329 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-dns-svc\") pod \"dnsmasq-dns-67b789f86c-6wwml\" (UID: \"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a\") " pod="openstack/dnsmasq-dns-67b789f86c-6wwml" Sep 30 14:16:46 crc kubenswrapper[4840]: I0930 14:16:46.089380 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-6wwml\" (UID: \"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a\") " pod="openstack/dnsmasq-dns-67b789f86c-6wwml" Sep 30 14:16:46 crc kubenswrapper[4840]: I0930 14:16:46.090006 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-dns-svc\") pod \"dnsmasq-dns-67b789f86c-6wwml\" (UID: \"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a\") " pod="openstack/dnsmasq-dns-67b789f86c-6wwml" Sep 30 14:16:46 crc kubenswrapper[4840]: I0930 14:16:46.090112 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-6wwml\" (UID: \"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a\") " pod="openstack/dnsmasq-dns-67b789f86c-6wwml" Sep 30 14:16:46 crc kubenswrapper[4840]: I0930 14:16:46.131048 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c15070b3-f247-4879-a7dc-618faf7e6e35" path="/var/lib/kubelet/pods/c15070b3-f247-4879-a7dc-618faf7e6e35/volumes" Sep 30 14:16:46 crc kubenswrapper[4840]: I0930 14:16:46.169154 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5z2dd\" (UniqueName: \"kubernetes.io/projected/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-kube-api-access-5z2dd\") pod \"dnsmasq-dns-67b789f86c-6wwml\" (UID: \"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a\") " pod="openstack/dnsmasq-dns-67b789f86c-6wwml" Sep 30 14:16:46 crc kubenswrapper[4840]: I0930 14:16:46.222119 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"533b8943-b9ea-4062-9dd2-ed1b29d4f2c6","Type":"ContainerStarted","Data":"5f30cbf205cf38c9fc8ed8873f962468954ffe9320b4aa942edc659385add719"} Sep 30 14:16:46 crc kubenswrapper[4840]: I0930 14:16:46.222171 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"533b8943-b9ea-4062-9dd2-ed1b29d4f2c6","Type":"ContainerStarted","Data":"685109a70fe1f2d4376acdd65ca2953996c834e09892aa99db30cb113f18c510"} Sep 30 14:16:46 crc kubenswrapper[4840]: I0930 14:16:46.265902 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-6wwml" Sep 30 14:16:46 crc kubenswrapper[4840]: I0930 14:16:46.746114 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-6wwml"] Sep 30 14:16:46 crc kubenswrapper[4840]: W0930 14:16:46.747371 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8aa6e768_9c67_43b6_a5ed_dfd06d4f4b6a.slice/crio-c913156d9e52b98c3559f5f1b612a660514d4bcbcb5009bdf7c0c5a7db63fcb1 WatchSource:0}: Error finding container c913156d9e52b98c3559f5f1b612a660514d4bcbcb5009bdf7c0c5a7db63fcb1: Status 404 returned error can't find the container with id c913156d9e52b98c3559f5f1b612a660514d4bcbcb5009bdf7c0c5a7db63fcb1 Sep 30 14:16:47 crc kubenswrapper[4840]: I0930 14:16:47.232710 4840 generic.go:334] "Generic (PLEG): container finished" podID="8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a" containerID="f3bd7a400d5c16eda5adeb21048cbcd0db5a01c3269145697d181940d30c82a8" exitCode=0 Sep 30 14:16:47 crc kubenswrapper[4840]: I0930 14:16:47.232813 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-6wwml" event={"ID":"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a","Type":"ContainerDied","Data":"f3bd7a400d5c16eda5adeb21048cbcd0db5a01c3269145697d181940d30c82a8"} Sep 30 14:16:47 crc kubenswrapper[4840]: I0930 14:16:47.233180 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-6wwml" event={"ID":"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a","Type":"ContainerStarted","Data":"c913156d9e52b98c3559f5f1b612a660514d4bcbcb5009bdf7c0c5a7db63fcb1"} Sep 30 14:16:47 crc kubenswrapper[4840]: I0930 14:16:47.453164 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Sep 30 14:16:47 crc kubenswrapper[4840]: I0930 14:16:47.479906 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Sep 30 14:16:48 crc kubenswrapper[4840]: I0930 14:16:48.243727 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-6wwml" event={"ID":"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a","Type":"ContainerStarted","Data":"34670ae962209e96c7c1b3443334092fb9633740ba1cb2182405740be339d96c"} Sep 30 14:16:48 crc kubenswrapper[4840]: I0930 14:16:48.272488 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67b789f86c-6wwml" podStartSLOduration=3.27246979 podStartE2EDuration="3.27246979s" podCreationTimestamp="2025-09-30 14:16:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:16:48.269965516 +0000 UTC m=+1236.899051939" watchObservedRunningTime="2025-09-30 14:16:48.27246979 +0000 UTC m=+1236.901556213" Sep 30 14:16:48 crc kubenswrapper[4840]: I0930 14:16:48.273177 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Sep 30 14:16:49 crc kubenswrapper[4840]: I0930 14:16:49.251648 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67b789f86c-6wwml" Sep 30 14:16:50 crc kubenswrapper[4840]: I0930 14:16:50.395236 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Sep 30 14:16:50 crc kubenswrapper[4840]: I0930 14:16:50.398196 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Sep 30 14:16:50 crc kubenswrapper[4840]: I0930 14:16:50.402012 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Sep 30 14:16:51 crc kubenswrapper[4840]: I0930 14:16:51.273215 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Sep 30 14:16:51 crc kubenswrapper[4840]: I0930 14:16:51.443795 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Sep 30 14:16:51 crc kubenswrapper[4840]: I0930 14:16:51.443864 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Sep 30 14:16:51 crc kubenswrapper[4840]: I0930 14:16:51.444461 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Sep 30 14:16:51 crc kubenswrapper[4840]: I0930 14:16:51.444497 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Sep 30 14:16:51 crc kubenswrapper[4840]: I0930 14:16:51.451675 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Sep 30 14:16:51 crc kubenswrapper[4840]: I0930 14:16:51.452221 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Sep 30 14:16:51 crc kubenswrapper[4840]: I0930 14:16:51.871652 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:16:51 crc kubenswrapper[4840]: I0930 14:16:51.871708 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:16:56 crc kubenswrapper[4840]: I0930 14:16:56.268155 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-67b789f86c-6wwml" Sep 30 14:16:56 crc kubenswrapper[4840]: I0930 14:16:56.332288 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-mxvtv"] Sep 30 14:16:56 crc kubenswrapper[4840]: I0930 14:16:56.332581 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59cf4bdb65-mxvtv" podUID="03083423-b558-40aa-afda-c29cef215e79" containerName="dnsmasq-dns" containerID="cri-o://145a9bfae9f1aa099ac4d9ad79e5402389fe0a558a6fd79bac70812085b8b451" gracePeriod=10 Sep 30 14:16:56 crc kubenswrapper[4840]: I0930 14:16:56.472297 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-87k2l"] Sep 30 14:16:56 crc kubenswrapper[4840]: I0930 14:16:56.474228 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb6ffcf87-87k2l" Sep 30 14:16:56 crc kubenswrapper[4840]: I0930 14:16:56.493692 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-87k2l"] Sep 30 14:16:56 crc kubenswrapper[4840]: I0930 14:16:56.588726 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/462dd9f7-ec69-4883-aa78-9b014c449baa-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-87k2l\" (UID: \"462dd9f7-ec69-4883-aa78-9b014c449baa\") " pod="openstack/dnsmasq-dns-cb6ffcf87-87k2l" Sep 30 14:16:56 crc kubenswrapper[4840]: I0930 14:16:56.588793 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/462dd9f7-ec69-4883-aa78-9b014c449baa-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-87k2l\" (UID: \"462dd9f7-ec69-4883-aa78-9b014c449baa\") " pod="openstack/dnsmasq-dns-cb6ffcf87-87k2l" Sep 30 14:16:56 crc kubenswrapper[4840]: I0930 14:16:56.588828 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gxjz\" (UniqueName: \"kubernetes.io/projected/462dd9f7-ec69-4883-aa78-9b014c449baa-kube-api-access-4gxjz\") pod \"dnsmasq-dns-cb6ffcf87-87k2l\" (UID: \"462dd9f7-ec69-4883-aa78-9b014c449baa\") " pod="openstack/dnsmasq-dns-cb6ffcf87-87k2l" Sep 30 14:16:56 crc kubenswrapper[4840]: I0930 14:16:56.588862 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/462dd9f7-ec69-4883-aa78-9b014c449baa-config\") pod \"dnsmasq-dns-cb6ffcf87-87k2l\" (UID: \"462dd9f7-ec69-4883-aa78-9b014c449baa\") " pod="openstack/dnsmasq-dns-cb6ffcf87-87k2l" Sep 30 14:16:56 crc kubenswrapper[4840]: I0930 14:16:56.588909 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/462dd9f7-ec69-4883-aa78-9b014c449baa-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-87k2l\" (UID: \"462dd9f7-ec69-4883-aa78-9b014c449baa\") " pod="openstack/dnsmasq-dns-cb6ffcf87-87k2l" Sep 30 14:16:56 crc kubenswrapper[4840]: I0930 14:16:56.588955 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/462dd9f7-ec69-4883-aa78-9b014c449baa-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-87k2l\" (UID: \"462dd9f7-ec69-4883-aa78-9b014c449baa\") " pod="openstack/dnsmasq-dns-cb6ffcf87-87k2l" Sep 30 14:16:56 crc kubenswrapper[4840]: I0930 14:16:56.588987 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/462dd9f7-ec69-4883-aa78-9b014c449baa-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-87k2l\" (UID: \"462dd9f7-ec69-4883-aa78-9b014c449baa\") " pod="openstack/dnsmasq-dns-cb6ffcf87-87k2l" Sep 30 14:16:56 crc kubenswrapper[4840]: I0930 14:16:56.690023 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/462dd9f7-ec69-4883-aa78-9b014c449baa-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-87k2l\" (UID: \"462dd9f7-ec69-4883-aa78-9b014c449baa\") " pod="openstack/dnsmasq-dns-cb6ffcf87-87k2l" Sep 30 14:16:56 crc kubenswrapper[4840]: I0930 14:16:56.690111 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/462dd9f7-ec69-4883-aa78-9b014c449baa-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-87k2l\" (UID: \"462dd9f7-ec69-4883-aa78-9b014c449baa\") " pod="openstack/dnsmasq-dns-cb6ffcf87-87k2l" Sep 30 14:16:56 crc kubenswrapper[4840]: I0930 14:16:56.690151 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/462dd9f7-ec69-4883-aa78-9b014c449baa-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-87k2l\" (UID: \"462dd9f7-ec69-4883-aa78-9b014c449baa\") " pod="openstack/dnsmasq-dns-cb6ffcf87-87k2l" Sep 30 14:16:56 crc kubenswrapper[4840]: I0930 14:16:56.690187 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/462dd9f7-ec69-4883-aa78-9b014c449baa-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-87k2l\" (UID: \"462dd9f7-ec69-4883-aa78-9b014c449baa\") " pod="openstack/dnsmasq-dns-cb6ffcf87-87k2l" Sep 30 14:16:56 crc kubenswrapper[4840]: I0930 14:16:56.690220 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/462dd9f7-ec69-4883-aa78-9b014c449baa-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-87k2l\" (UID: \"462dd9f7-ec69-4883-aa78-9b014c449baa\") " pod="openstack/dnsmasq-dns-cb6ffcf87-87k2l" Sep 30 14:16:56 crc kubenswrapper[4840]: I0930 14:16:56.690249 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gxjz\" (UniqueName: \"kubernetes.io/projected/462dd9f7-ec69-4883-aa78-9b014c449baa-kube-api-access-4gxjz\") pod \"dnsmasq-dns-cb6ffcf87-87k2l\" (UID: \"462dd9f7-ec69-4883-aa78-9b014c449baa\") " pod="openstack/dnsmasq-dns-cb6ffcf87-87k2l" Sep 30 14:16:56 crc kubenswrapper[4840]: I0930 14:16:56.690279 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/462dd9f7-ec69-4883-aa78-9b014c449baa-config\") pod \"dnsmasq-dns-cb6ffcf87-87k2l\" (UID: \"462dd9f7-ec69-4883-aa78-9b014c449baa\") " pod="openstack/dnsmasq-dns-cb6ffcf87-87k2l" Sep 30 14:16:56 crc kubenswrapper[4840]: I0930 14:16:56.691076 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/462dd9f7-ec69-4883-aa78-9b014c449baa-config\") pod \"dnsmasq-dns-cb6ffcf87-87k2l\" (UID: \"462dd9f7-ec69-4883-aa78-9b014c449baa\") " pod="openstack/dnsmasq-dns-cb6ffcf87-87k2l" Sep 30 14:16:56 crc kubenswrapper[4840]: I0930 14:16:56.691693 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/462dd9f7-ec69-4883-aa78-9b014c449baa-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-87k2l\" (UID: \"462dd9f7-ec69-4883-aa78-9b014c449baa\") " pod="openstack/dnsmasq-dns-cb6ffcf87-87k2l" Sep 30 14:16:56 crc kubenswrapper[4840]: I0930 14:16:56.692265 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/462dd9f7-ec69-4883-aa78-9b014c449baa-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-87k2l\" (UID: \"462dd9f7-ec69-4883-aa78-9b014c449baa\") " pod="openstack/dnsmasq-dns-cb6ffcf87-87k2l" Sep 30 14:16:56 crc kubenswrapper[4840]: I0930 14:16:56.693009 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/462dd9f7-ec69-4883-aa78-9b014c449baa-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-87k2l\" (UID: \"462dd9f7-ec69-4883-aa78-9b014c449baa\") " pod="openstack/dnsmasq-dns-cb6ffcf87-87k2l" Sep 30 14:16:56 crc kubenswrapper[4840]: I0930 14:16:56.693423 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/462dd9f7-ec69-4883-aa78-9b014c449baa-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-87k2l\" (UID: \"462dd9f7-ec69-4883-aa78-9b014c449baa\") " pod="openstack/dnsmasq-dns-cb6ffcf87-87k2l" Sep 30 14:16:56 crc kubenswrapper[4840]: I0930 14:16:56.693423 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/462dd9f7-ec69-4883-aa78-9b014c449baa-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-87k2l\" (UID: \"462dd9f7-ec69-4883-aa78-9b014c449baa\") " pod="openstack/dnsmasq-dns-cb6ffcf87-87k2l" Sep 30 14:16:56 crc kubenswrapper[4840]: I0930 14:16:56.712991 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gxjz\" (UniqueName: \"kubernetes.io/projected/462dd9f7-ec69-4883-aa78-9b014c449baa-kube-api-access-4gxjz\") pod \"dnsmasq-dns-cb6ffcf87-87k2l\" (UID: \"462dd9f7-ec69-4883-aa78-9b014c449baa\") " pod="openstack/dnsmasq-dns-cb6ffcf87-87k2l" Sep 30 14:16:56 crc kubenswrapper[4840]: I0930 14:16:56.817648 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-mxvtv" Sep 30 14:16:56 crc kubenswrapper[4840]: I0930 14:16:56.821533 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb6ffcf87-87k2l" Sep 30 14:16:56 crc kubenswrapper[4840]: I0930 14:16:56.997307 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/03083423-b558-40aa-afda-c29cef215e79-ovsdbserver-sb\") pod \"03083423-b558-40aa-afda-c29cef215e79\" (UID: \"03083423-b558-40aa-afda-c29cef215e79\") " Sep 30 14:16:56 crc kubenswrapper[4840]: I0930 14:16:56.997347 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/03083423-b558-40aa-afda-c29cef215e79-dns-swift-storage-0\") pod \"03083423-b558-40aa-afda-c29cef215e79\" (UID: \"03083423-b558-40aa-afda-c29cef215e79\") " Sep 30 14:16:56 crc kubenswrapper[4840]: I0930 14:16:56.997455 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03083423-b558-40aa-afda-c29cef215e79-config\") pod \"03083423-b558-40aa-afda-c29cef215e79\" (UID: \"03083423-b558-40aa-afda-c29cef215e79\") " Sep 30 14:16:56 crc kubenswrapper[4840]: I0930 14:16:56.997489 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5kvj\" (UniqueName: \"kubernetes.io/projected/03083423-b558-40aa-afda-c29cef215e79-kube-api-access-j5kvj\") pod \"03083423-b558-40aa-afda-c29cef215e79\" (UID: \"03083423-b558-40aa-afda-c29cef215e79\") " Sep 30 14:16:56 crc kubenswrapper[4840]: I0930 14:16:56.997534 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/03083423-b558-40aa-afda-c29cef215e79-ovsdbserver-nb\") pod \"03083423-b558-40aa-afda-c29cef215e79\" (UID: \"03083423-b558-40aa-afda-c29cef215e79\") " Sep 30 14:16:56 crc kubenswrapper[4840]: I0930 14:16:56.997606 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03083423-b558-40aa-afda-c29cef215e79-dns-svc\") pod \"03083423-b558-40aa-afda-c29cef215e79\" (UID: \"03083423-b558-40aa-afda-c29cef215e79\") " Sep 30 14:16:57 crc kubenswrapper[4840]: I0930 14:16:57.014015 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03083423-b558-40aa-afda-c29cef215e79-kube-api-access-j5kvj" (OuterVolumeSpecName: "kube-api-access-j5kvj") pod "03083423-b558-40aa-afda-c29cef215e79" (UID: "03083423-b558-40aa-afda-c29cef215e79"). InnerVolumeSpecName "kube-api-access-j5kvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:16:57 crc kubenswrapper[4840]: I0930 14:16:57.054371 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03083423-b558-40aa-afda-c29cef215e79-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "03083423-b558-40aa-afda-c29cef215e79" (UID: "03083423-b558-40aa-afda-c29cef215e79"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:16:57 crc kubenswrapper[4840]: I0930 14:16:57.061013 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03083423-b558-40aa-afda-c29cef215e79-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "03083423-b558-40aa-afda-c29cef215e79" (UID: "03083423-b558-40aa-afda-c29cef215e79"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:16:57 crc kubenswrapper[4840]: I0930 14:16:57.065067 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03083423-b558-40aa-afda-c29cef215e79-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "03083423-b558-40aa-afda-c29cef215e79" (UID: "03083423-b558-40aa-afda-c29cef215e79"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:16:57 crc kubenswrapper[4840]: I0930 14:16:57.066669 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03083423-b558-40aa-afda-c29cef215e79-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "03083423-b558-40aa-afda-c29cef215e79" (UID: "03083423-b558-40aa-afda-c29cef215e79"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:16:57 crc kubenswrapper[4840]: I0930 14:16:57.068565 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03083423-b558-40aa-afda-c29cef215e79-config" (OuterVolumeSpecName: "config") pod "03083423-b558-40aa-afda-c29cef215e79" (UID: "03083423-b558-40aa-afda-c29cef215e79"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:16:57 crc kubenswrapper[4840]: I0930 14:16:57.100238 4840 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/03083423-b558-40aa-afda-c29cef215e79-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:57 crc kubenswrapper[4840]: I0930 14:16:57.100266 4840 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/03083423-b558-40aa-afda-c29cef215e79-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:57 crc kubenswrapper[4840]: I0930 14:16:57.100278 4840 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03083423-b558-40aa-afda-c29cef215e79-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:57 crc kubenswrapper[4840]: I0930 14:16:57.100286 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5kvj\" (UniqueName: \"kubernetes.io/projected/03083423-b558-40aa-afda-c29cef215e79-kube-api-access-j5kvj\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:57 crc kubenswrapper[4840]: I0930 14:16:57.100298 4840 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/03083423-b558-40aa-afda-c29cef215e79-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:57 crc kubenswrapper[4840]: I0930 14:16:57.100307 4840 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03083423-b558-40aa-afda-c29cef215e79-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 14:16:57 crc kubenswrapper[4840]: I0930 14:16:57.303379 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-87k2l"] Sep 30 14:16:57 crc kubenswrapper[4840]: I0930 14:16:57.327544 4840 generic.go:334] "Generic (PLEG): container finished" podID="03083423-b558-40aa-afda-c29cef215e79" containerID="145a9bfae9f1aa099ac4d9ad79e5402389fe0a558a6fd79bac70812085b8b451" exitCode=0 Sep 30 14:16:57 crc kubenswrapper[4840]: I0930 14:16:57.327648 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-mxvtv" event={"ID":"03083423-b558-40aa-afda-c29cef215e79","Type":"ContainerDied","Data":"145a9bfae9f1aa099ac4d9ad79e5402389fe0a558a6fd79bac70812085b8b451"} Sep 30 14:16:57 crc kubenswrapper[4840]: I0930 14:16:57.327674 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-mxvtv" event={"ID":"03083423-b558-40aa-afda-c29cef215e79","Type":"ContainerDied","Data":"0dfb69949186453edc5ef7345eecf59f56b9451b05f5fac3ac24e33abf7fce2d"} Sep 30 14:16:57 crc kubenswrapper[4840]: I0930 14:16:57.327691 4840 scope.go:117] "RemoveContainer" containerID="145a9bfae9f1aa099ac4d9ad79e5402389fe0a558a6fd79bac70812085b8b451" Sep 30 14:16:57 crc kubenswrapper[4840]: I0930 14:16:57.327816 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-mxvtv" Sep 30 14:16:57 crc kubenswrapper[4840]: I0930 14:16:57.331005 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-87k2l" event={"ID":"462dd9f7-ec69-4883-aa78-9b014c449baa","Type":"ContainerStarted","Data":"1792274653a874a95c5fecee7a4de09c543d3391c51483cf7fb45c0eb6f2c0d5"} Sep 30 14:16:57 crc kubenswrapper[4840]: I0930 14:16:57.347782 4840 scope.go:117] "RemoveContainer" containerID="952a44017306d2fd9b1f12d571014db74c53629ec6dfbb16338203af55cf8d74" Sep 30 14:16:57 crc kubenswrapper[4840]: I0930 14:16:57.361239 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-mxvtv"] Sep 30 14:16:57 crc kubenswrapper[4840]: I0930 14:16:57.370381 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-mxvtv"] Sep 30 14:16:57 crc kubenswrapper[4840]: I0930 14:16:57.377939 4840 scope.go:117] "RemoveContainer" containerID="145a9bfae9f1aa099ac4d9ad79e5402389fe0a558a6fd79bac70812085b8b451" Sep 30 14:16:57 crc kubenswrapper[4840]: E0930 14:16:57.378449 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"145a9bfae9f1aa099ac4d9ad79e5402389fe0a558a6fd79bac70812085b8b451\": container with ID starting with 145a9bfae9f1aa099ac4d9ad79e5402389fe0a558a6fd79bac70812085b8b451 not found: ID does not exist" containerID="145a9bfae9f1aa099ac4d9ad79e5402389fe0a558a6fd79bac70812085b8b451" Sep 30 14:16:57 crc kubenswrapper[4840]: I0930 14:16:57.378488 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"145a9bfae9f1aa099ac4d9ad79e5402389fe0a558a6fd79bac70812085b8b451"} err="failed to get container status \"145a9bfae9f1aa099ac4d9ad79e5402389fe0a558a6fd79bac70812085b8b451\": rpc error: code = NotFound desc = could not find container \"145a9bfae9f1aa099ac4d9ad79e5402389fe0a558a6fd79bac70812085b8b451\": container with ID starting with 145a9bfae9f1aa099ac4d9ad79e5402389fe0a558a6fd79bac70812085b8b451 not found: ID does not exist" Sep 30 14:16:57 crc kubenswrapper[4840]: I0930 14:16:57.378512 4840 scope.go:117] "RemoveContainer" containerID="952a44017306d2fd9b1f12d571014db74c53629ec6dfbb16338203af55cf8d74" Sep 30 14:16:57 crc kubenswrapper[4840]: E0930 14:16:57.378931 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"952a44017306d2fd9b1f12d571014db74c53629ec6dfbb16338203af55cf8d74\": container with ID starting with 952a44017306d2fd9b1f12d571014db74c53629ec6dfbb16338203af55cf8d74 not found: ID does not exist" containerID="952a44017306d2fd9b1f12d571014db74c53629ec6dfbb16338203af55cf8d74" Sep 30 14:16:57 crc kubenswrapper[4840]: I0930 14:16:57.378965 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"952a44017306d2fd9b1f12d571014db74c53629ec6dfbb16338203af55cf8d74"} err="failed to get container status \"952a44017306d2fd9b1f12d571014db74c53629ec6dfbb16338203af55cf8d74\": rpc error: code = NotFound desc = could not find container \"952a44017306d2fd9b1f12d571014db74c53629ec6dfbb16338203af55cf8d74\": container with ID starting with 952a44017306d2fd9b1f12d571014db74c53629ec6dfbb16338203af55cf8d74 not found: ID does not exist" Sep 30 14:16:58 crc kubenswrapper[4840]: I0930 14:16:58.127267 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03083423-b558-40aa-afda-c29cef215e79" path="/var/lib/kubelet/pods/03083423-b558-40aa-afda-c29cef215e79/volumes" Sep 30 14:16:58 crc kubenswrapper[4840]: I0930 14:16:58.340318 4840 generic.go:334] "Generic (PLEG): container finished" podID="462dd9f7-ec69-4883-aa78-9b014c449baa" containerID="b35d6002a0e449fb18df9bd5e909ae281c417420b0076765abe2c5f8e8bb4227" exitCode=0 Sep 30 14:16:58 crc kubenswrapper[4840]: I0930 14:16:58.340379 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-87k2l" event={"ID":"462dd9f7-ec69-4883-aa78-9b014c449baa","Type":"ContainerDied","Data":"b35d6002a0e449fb18df9bd5e909ae281c417420b0076765abe2c5f8e8bb4227"} Sep 30 14:16:59 crc kubenswrapper[4840]: I0930 14:16:59.353990 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-87k2l" event={"ID":"462dd9f7-ec69-4883-aa78-9b014c449baa","Type":"ContainerStarted","Data":"dd72b27ae4e531d120a521e8dcf4f35d83faa84e2b11481f9b309606ca0582ba"} Sep 30 14:16:59 crc kubenswrapper[4840]: I0930 14:16:59.354261 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cb6ffcf87-87k2l" Sep 30 14:16:59 crc kubenswrapper[4840]: I0930 14:16:59.378143 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cb6ffcf87-87k2l" podStartSLOduration=3.378122728 podStartE2EDuration="3.378122728s" podCreationTimestamp="2025-09-30 14:16:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:16:59.37390966 +0000 UTC m=+1248.002996093" watchObservedRunningTime="2025-09-30 14:16:59.378122728 +0000 UTC m=+1248.007209151" Sep 30 14:17:06 crc kubenswrapper[4840]: I0930 14:17:06.823506 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cb6ffcf87-87k2l" Sep 30 14:17:06 crc kubenswrapper[4840]: I0930 14:17:06.922144 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-6wwml"] Sep 30 14:17:06 crc kubenswrapper[4840]: I0930 14:17:06.922661 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-67b789f86c-6wwml" podUID="8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a" containerName="dnsmasq-dns" containerID="cri-o://34670ae962209e96c7c1b3443334092fb9633740ba1cb2182405740be339d96c" gracePeriod=10 Sep 30 14:17:07 crc kubenswrapper[4840]: I0930 14:17:07.381156 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-6wwml" Sep 30 14:17:07 crc kubenswrapper[4840]: I0930 14:17:07.438961 4840 generic.go:334] "Generic (PLEG): container finished" podID="8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a" containerID="34670ae962209e96c7c1b3443334092fb9633740ba1cb2182405740be339d96c" exitCode=0 Sep 30 14:17:07 crc kubenswrapper[4840]: I0930 14:17:07.439030 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-6wwml" Sep 30 14:17:07 crc kubenswrapper[4840]: I0930 14:17:07.439015 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-6wwml" event={"ID":"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a","Type":"ContainerDied","Data":"34670ae962209e96c7c1b3443334092fb9633740ba1cb2182405740be339d96c"} Sep 30 14:17:07 crc kubenswrapper[4840]: I0930 14:17:07.439172 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-6wwml" event={"ID":"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a","Type":"ContainerDied","Data":"c913156d9e52b98c3559f5f1b612a660514d4bcbcb5009bdf7c0c5a7db63fcb1"} Sep 30 14:17:07 crc kubenswrapper[4840]: I0930 14:17:07.439193 4840 scope.go:117] "RemoveContainer" containerID="34670ae962209e96c7c1b3443334092fb9633740ba1cb2182405740be339d96c" Sep 30 14:17:07 crc kubenswrapper[4840]: I0930 14:17:07.460390 4840 scope.go:117] "RemoveContainer" containerID="f3bd7a400d5c16eda5adeb21048cbcd0db5a01c3269145697d181940d30c82a8" Sep 30 14:17:07 crc kubenswrapper[4840]: I0930 14:17:07.485010 4840 scope.go:117] "RemoveContainer" containerID="34670ae962209e96c7c1b3443334092fb9633740ba1cb2182405740be339d96c" Sep 30 14:17:07 crc kubenswrapper[4840]: E0930 14:17:07.486481 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34670ae962209e96c7c1b3443334092fb9633740ba1cb2182405740be339d96c\": container with ID starting with 34670ae962209e96c7c1b3443334092fb9633740ba1cb2182405740be339d96c not found: ID does not exist" containerID="34670ae962209e96c7c1b3443334092fb9633740ba1cb2182405740be339d96c" Sep 30 14:17:07 crc kubenswrapper[4840]: I0930 14:17:07.486532 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34670ae962209e96c7c1b3443334092fb9633740ba1cb2182405740be339d96c"} err="failed to get container status \"34670ae962209e96c7c1b3443334092fb9633740ba1cb2182405740be339d96c\": rpc error: code = NotFound desc = could not find container \"34670ae962209e96c7c1b3443334092fb9633740ba1cb2182405740be339d96c\": container with ID starting with 34670ae962209e96c7c1b3443334092fb9633740ba1cb2182405740be339d96c not found: ID does not exist" Sep 30 14:17:07 crc kubenswrapper[4840]: I0930 14:17:07.486578 4840 scope.go:117] "RemoveContainer" containerID="f3bd7a400d5c16eda5adeb21048cbcd0db5a01c3269145697d181940d30c82a8" Sep 30 14:17:07 crc kubenswrapper[4840]: E0930 14:17:07.486920 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3bd7a400d5c16eda5adeb21048cbcd0db5a01c3269145697d181940d30c82a8\": container with ID starting with f3bd7a400d5c16eda5adeb21048cbcd0db5a01c3269145697d181940d30c82a8 not found: ID does not exist" containerID="f3bd7a400d5c16eda5adeb21048cbcd0db5a01c3269145697d181940d30c82a8" Sep 30 14:17:07 crc kubenswrapper[4840]: I0930 14:17:07.486947 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3bd7a400d5c16eda5adeb21048cbcd0db5a01c3269145697d181940d30c82a8"} err="failed to get container status \"f3bd7a400d5c16eda5adeb21048cbcd0db5a01c3269145697d181940d30c82a8\": rpc error: code = NotFound desc = could not find container \"f3bd7a400d5c16eda5adeb21048cbcd0db5a01c3269145697d181940d30c82a8\": container with ID starting with f3bd7a400d5c16eda5adeb21048cbcd0db5a01c3269145697d181940d30c82a8 not found: ID does not exist" Sep 30 14:17:07 crc kubenswrapper[4840]: I0930 14:17:07.511998 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-openstack-edpm-ipam\") pod \"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a\" (UID: \"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a\") " Sep 30 14:17:07 crc kubenswrapper[4840]: I0930 14:17:07.512061 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-ovsdbserver-nb\") pod \"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a\" (UID: \"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a\") " Sep 30 14:17:07 crc kubenswrapper[4840]: I0930 14:17:07.512080 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-config\") pod \"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a\" (UID: \"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a\") " Sep 30 14:17:07 crc kubenswrapper[4840]: I0930 14:17:07.512145 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-dns-swift-storage-0\") pod \"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a\" (UID: \"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a\") " Sep 30 14:17:07 crc kubenswrapper[4840]: I0930 14:17:07.512271 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-dns-svc\") pod \"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a\" (UID: \"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a\") " Sep 30 14:17:07 crc kubenswrapper[4840]: I0930 14:17:07.512356 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-ovsdbserver-sb\") pod \"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a\" (UID: \"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a\") " Sep 30 14:17:07 crc kubenswrapper[4840]: I0930 14:17:07.512401 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5z2dd\" (UniqueName: \"kubernetes.io/projected/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-kube-api-access-5z2dd\") pod \"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a\" (UID: \"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a\") " Sep 30 14:17:07 crc kubenswrapper[4840]: I0930 14:17:07.519139 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-kube-api-access-5z2dd" (OuterVolumeSpecName: "kube-api-access-5z2dd") pod "8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a" (UID: "8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a"). InnerVolumeSpecName "kube-api-access-5z2dd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:17:07 crc kubenswrapper[4840]: I0930 14:17:07.565655 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a" (UID: "8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:17:07 crc kubenswrapper[4840]: I0930 14:17:07.567903 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a" (UID: "8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:17:07 crc kubenswrapper[4840]: I0930 14:17:07.572371 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a" (UID: "8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:17:07 crc kubenswrapper[4840]: I0930 14:17:07.577157 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-config" (OuterVolumeSpecName: "config") pod "8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a" (UID: "8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:17:07 crc kubenswrapper[4840]: E0930 14:17:07.587203 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-dns-swift-storage-0 podName:8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a nodeName:}" failed. No retries permitted until 2025-09-30 14:17:08.087165907 +0000 UTC m=+1256.716252330 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "dns-swift-storage-0" (UniqueName: "kubernetes.io/configmap/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-dns-swift-storage-0") pod "8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a" (UID: "8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a") : error deleting /var/lib/kubelet/pods/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a/volume-subpaths: remove /var/lib/kubelet/pods/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a/volume-subpaths: no such file or directory Sep 30 14:17:07 crc kubenswrapper[4840]: I0930 14:17:07.587487 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a" (UID: "8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:17:07 crc kubenswrapper[4840]: I0930 14:17:07.614591 4840 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 14:17:07 crc kubenswrapper[4840]: I0930 14:17:07.614631 4840 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 14:17:07 crc kubenswrapper[4840]: I0930 14:17:07.614645 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5z2dd\" (UniqueName: \"kubernetes.io/projected/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-kube-api-access-5z2dd\") on node \"crc\" DevicePath \"\"" Sep 30 14:17:07 crc kubenswrapper[4840]: I0930 14:17:07.614655 4840 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Sep 30 14:17:07 crc kubenswrapper[4840]: I0930 14:17:07.614663 4840 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 14:17:07 crc kubenswrapper[4840]: I0930 14:17:07.614673 4840 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:17:08 crc kubenswrapper[4840]: I0930 14:17:08.123003 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-dns-swift-storage-0\") pod \"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a\" (UID: \"8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a\") " Sep 30 14:17:08 crc kubenswrapper[4840]: I0930 14:17:08.123865 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a" (UID: "8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:17:08 crc kubenswrapper[4840]: I0930 14:17:08.225273 4840 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:17:08 crc kubenswrapper[4840]: I0930 14:17:08.366309 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-6wwml"] Sep 30 14:17:08 crc kubenswrapper[4840]: I0930 14:17:08.374013 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-6wwml"] Sep 30 14:17:10 crc kubenswrapper[4840]: I0930 14:17:10.132927 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a" path="/var/lib/kubelet/pods/8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a/volumes" Sep 30 14:17:15 crc kubenswrapper[4840]: I0930 14:17:15.517148 4840 generic.go:334] "Generic (PLEG): container finished" podID="97b46089-3117-4a2a-b9db-ada9d0daca8b" containerID="dcee92a5252d73c0d68f435241b274fd172504116b62bc26d71ec510c69e33a5" exitCode=0 Sep 30 14:17:15 crc kubenswrapper[4840]: I0930 14:17:15.517314 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"97b46089-3117-4a2a-b9db-ada9d0daca8b","Type":"ContainerDied","Data":"dcee92a5252d73c0d68f435241b274fd172504116b62bc26d71ec510c69e33a5"} Sep 30 14:17:16 crc kubenswrapper[4840]: I0930 14:17:16.532456 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"97b46089-3117-4a2a-b9db-ada9d0daca8b","Type":"ContainerStarted","Data":"f8798f29a3929e9082c3f1b2ef4369e2c898456ceef05a7f3ff3e12dfdcf02fa"} Sep 30 14:17:16 crc kubenswrapper[4840]: I0930 14:17:16.533036 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Sep 30 14:17:16 crc kubenswrapper[4840]: I0930 14:17:16.534716 4840 generic.go:334] "Generic (PLEG): container finished" podID="533b8943-b9ea-4062-9dd2-ed1b29d4f2c6" containerID="5f30cbf205cf38c9fc8ed8873f962468954ffe9320b4aa942edc659385add719" exitCode=0 Sep 30 14:17:16 crc kubenswrapper[4840]: I0930 14:17:16.534741 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"533b8943-b9ea-4062-9dd2-ed1b29d4f2c6","Type":"ContainerDied","Data":"5f30cbf205cf38c9fc8ed8873f962468954ffe9320b4aa942edc659385add719"} Sep 30 14:17:16 crc kubenswrapper[4840]: I0930 14:17:16.569162 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=33.569147143 podStartE2EDuration="33.569147143s" podCreationTimestamp="2025-09-30 14:16:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:17:16.56514036 +0000 UTC m=+1265.194226863" watchObservedRunningTime="2025-09-30 14:17:16.569147143 +0000 UTC m=+1265.198233566" Sep 30 14:17:17 crc kubenswrapper[4840]: I0930 14:17:17.546427 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"533b8943-b9ea-4062-9dd2-ed1b29d4f2c6","Type":"ContainerStarted","Data":"4b237fcac9e81b3764fa4a29c4eb66521046e811ced93d49e11201c17f662b79"} Sep 30 14:17:17 crc kubenswrapper[4840]: I0930 14:17:17.547640 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:17:17 crc kubenswrapper[4840]: I0930 14:17:17.584581 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=33.584543781 podStartE2EDuration="33.584543781s" podCreationTimestamp="2025-09-30 14:16:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:17:17.578045314 +0000 UTC m=+1266.207131737" watchObservedRunningTime="2025-09-30 14:17:17.584543781 +0000 UTC m=+1266.213630204" Sep 30 14:17:20 crc kubenswrapper[4840]: I0930 14:17:20.184094 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q5t88"] Sep 30 14:17:20 crc kubenswrapper[4840]: E0930 14:17:20.184575 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03083423-b558-40aa-afda-c29cef215e79" containerName="dnsmasq-dns" Sep 30 14:17:20 crc kubenswrapper[4840]: I0930 14:17:20.184594 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="03083423-b558-40aa-afda-c29cef215e79" containerName="dnsmasq-dns" Sep 30 14:17:20 crc kubenswrapper[4840]: E0930 14:17:20.184620 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a" containerName="init" Sep 30 14:17:20 crc kubenswrapper[4840]: I0930 14:17:20.184629 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a" containerName="init" Sep 30 14:17:20 crc kubenswrapper[4840]: E0930 14:17:20.184648 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03083423-b558-40aa-afda-c29cef215e79" containerName="init" Sep 30 14:17:20 crc kubenswrapper[4840]: I0930 14:17:20.184656 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="03083423-b558-40aa-afda-c29cef215e79" containerName="init" Sep 30 14:17:20 crc kubenswrapper[4840]: E0930 14:17:20.184685 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a" containerName="dnsmasq-dns" Sep 30 14:17:20 crc kubenswrapper[4840]: I0930 14:17:20.184694 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a" containerName="dnsmasq-dns" Sep 30 14:17:20 crc kubenswrapper[4840]: I0930 14:17:20.184948 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="03083423-b558-40aa-afda-c29cef215e79" containerName="dnsmasq-dns" Sep 30 14:17:20 crc kubenswrapper[4840]: I0930 14:17:20.184970 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="8aa6e768-9c67-43b6-a5ed-dfd06d4f4b6a" containerName="dnsmasq-dns" Sep 30 14:17:20 crc kubenswrapper[4840]: I0930 14:17:20.185771 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q5t88" Sep 30 14:17:20 crc kubenswrapper[4840]: I0930 14:17:20.190250 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 14:17:20 crc kubenswrapper[4840]: I0930 14:17:20.190516 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 14:17:20 crc kubenswrapper[4840]: I0930 14:17:20.191572 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 14:17:20 crc kubenswrapper[4840]: I0930 14:17:20.196223 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q5t88"] Sep 30 14:17:20 crc kubenswrapper[4840]: I0930 14:17:20.199970 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9zm9z" Sep 30 14:17:20 crc kubenswrapper[4840]: I0930 14:17:20.269014 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1073e6a-94e1-4fad-a116-ff4e145a0d22-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q5t88\" (UID: \"c1073e6a-94e1-4fad-a116-ff4e145a0d22\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q5t88" Sep 30 14:17:20 crc kubenswrapper[4840]: I0930 14:17:20.269087 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c1073e6a-94e1-4fad-a116-ff4e145a0d22-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q5t88\" (UID: \"c1073e6a-94e1-4fad-a116-ff4e145a0d22\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q5t88" Sep 30 14:17:20 crc kubenswrapper[4840]: I0930 14:17:20.269157 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rj4c\" (UniqueName: \"kubernetes.io/projected/c1073e6a-94e1-4fad-a116-ff4e145a0d22-kube-api-access-8rj4c\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q5t88\" (UID: \"c1073e6a-94e1-4fad-a116-ff4e145a0d22\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q5t88" Sep 30 14:17:20 crc kubenswrapper[4840]: I0930 14:17:20.269236 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c1073e6a-94e1-4fad-a116-ff4e145a0d22-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q5t88\" (UID: \"c1073e6a-94e1-4fad-a116-ff4e145a0d22\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q5t88" Sep 30 14:17:20 crc kubenswrapper[4840]: I0930 14:17:20.371499 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1073e6a-94e1-4fad-a116-ff4e145a0d22-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q5t88\" (UID: \"c1073e6a-94e1-4fad-a116-ff4e145a0d22\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q5t88" Sep 30 14:17:20 crc kubenswrapper[4840]: I0930 14:17:20.371608 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c1073e6a-94e1-4fad-a116-ff4e145a0d22-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q5t88\" (UID: \"c1073e6a-94e1-4fad-a116-ff4e145a0d22\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q5t88" Sep 30 14:17:20 crc kubenswrapper[4840]: I0930 14:17:20.371702 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rj4c\" (UniqueName: \"kubernetes.io/projected/c1073e6a-94e1-4fad-a116-ff4e145a0d22-kube-api-access-8rj4c\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q5t88\" (UID: \"c1073e6a-94e1-4fad-a116-ff4e145a0d22\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q5t88" Sep 30 14:17:20 crc kubenswrapper[4840]: I0930 14:17:20.371793 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c1073e6a-94e1-4fad-a116-ff4e145a0d22-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q5t88\" (UID: \"c1073e6a-94e1-4fad-a116-ff4e145a0d22\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q5t88" Sep 30 14:17:20 crc kubenswrapper[4840]: I0930 14:17:20.377930 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1073e6a-94e1-4fad-a116-ff4e145a0d22-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q5t88\" (UID: \"c1073e6a-94e1-4fad-a116-ff4e145a0d22\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q5t88" Sep 30 14:17:20 crc kubenswrapper[4840]: I0930 14:17:20.378260 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c1073e6a-94e1-4fad-a116-ff4e145a0d22-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q5t88\" (UID: \"c1073e6a-94e1-4fad-a116-ff4e145a0d22\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q5t88" Sep 30 14:17:20 crc kubenswrapper[4840]: I0930 14:17:20.378742 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c1073e6a-94e1-4fad-a116-ff4e145a0d22-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q5t88\" (UID: \"c1073e6a-94e1-4fad-a116-ff4e145a0d22\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q5t88" Sep 30 14:17:20 crc kubenswrapper[4840]: I0930 14:17:20.388924 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rj4c\" (UniqueName: \"kubernetes.io/projected/c1073e6a-94e1-4fad-a116-ff4e145a0d22-kube-api-access-8rj4c\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q5t88\" (UID: \"c1073e6a-94e1-4fad-a116-ff4e145a0d22\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q5t88" Sep 30 14:17:20 crc kubenswrapper[4840]: I0930 14:17:20.511054 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q5t88" Sep 30 14:17:21 crc kubenswrapper[4840]: I0930 14:17:21.053572 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q5t88"] Sep 30 14:17:21 crc kubenswrapper[4840]: I0930 14:17:21.067320 4840 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 14:17:21 crc kubenswrapper[4840]: I0930 14:17:21.582046 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q5t88" event={"ID":"c1073e6a-94e1-4fad-a116-ff4e145a0d22","Type":"ContainerStarted","Data":"6eb69b7b7435ecee029914f1292c73de123d03c9dfb482ba52fa14887e09fec5"} Sep 30 14:17:21 crc kubenswrapper[4840]: I0930 14:17:21.871390 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:17:21 crc kubenswrapper[4840]: I0930 14:17:21.871447 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:17:21 crc kubenswrapper[4840]: I0930 14:17:21.871488 4840 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-747gk" Sep 30 14:17:21 crc kubenswrapper[4840]: I0930 14:17:21.872211 4840 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5170b28c065e5cc00d1164411d96a8eb01e1860be6b5701def2ab76e6fb63860"} pod="openshift-machine-config-operator/machine-config-daemon-747gk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 14:17:21 crc kubenswrapper[4840]: I0930 14:17:21.872268 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" containerID="cri-o://5170b28c065e5cc00d1164411d96a8eb01e1860be6b5701def2ab76e6fb63860" gracePeriod=600 Sep 30 14:17:22 crc kubenswrapper[4840]: I0930 14:17:22.598352 4840 generic.go:334] "Generic (PLEG): container finished" podID="10e8b890-7f20-4a36-8e03-898620cf599a" containerID="5170b28c065e5cc00d1164411d96a8eb01e1860be6b5701def2ab76e6fb63860" exitCode=0 Sep 30 14:17:22 crc kubenswrapper[4840]: I0930 14:17:22.598387 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" event={"ID":"10e8b890-7f20-4a36-8e03-898620cf599a","Type":"ContainerDied","Data":"5170b28c065e5cc00d1164411d96a8eb01e1860be6b5701def2ab76e6fb63860"} Sep 30 14:17:22 crc kubenswrapper[4840]: I0930 14:17:22.599120 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" event={"ID":"10e8b890-7f20-4a36-8e03-898620cf599a","Type":"ContainerStarted","Data":"b4ec147091f38f805d294a67b19870b478b6bf3d54869eabd43f91f49541a76e"} Sep 30 14:17:22 crc kubenswrapper[4840]: I0930 14:17:22.599162 4840 scope.go:117] "RemoveContainer" containerID="6060b9705d9c27f06e58cff4fd0a3ee141a9c798fa228092898474533d2401f1" Sep 30 14:17:31 crc kubenswrapper[4840]: I0930 14:17:31.681391 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q5t88" event={"ID":"c1073e6a-94e1-4fad-a116-ff4e145a0d22","Type":"ContainerStarted","Data":"1d44e53a0dc10a4da37ddedd8788029adf23a290f62e570cbc12812dad290363"} Sep 30 14:17:31 crc kubenswrapper[4840]: I0930 14:17:31.710479 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q5t88" podStartSLOduration=1.49388557 podStartE2EDuration="11.710455735s" podCreationTimestamp="2025-09-30 14:17:20 +0000 UTC" firstStartedPulling="2025-09-30 14:17:21.067121916 +0000 UTC m=+1269.696208339" lastFinishedPulling="2025-09-30 14:17:31.283692081 +0000 UTC m=+1279.912778504" observedRunningTime="2025-09-30 14:17:31.698116128 +0000 UTC m=+1280.327202571" watchObservedRunningTime="2025-09-30 14:17:31.710455735 +0000 UTC m=+1280.339542168" Sep 30 14:17:32 crc kubenswrapper[4840]: I0930 14:17:32.735101 4840 scope.go:117] "RemoveContainer" containerID="b6817e74a280d4642bbb241ea913659cb63427fbe679a334fe8dfdf5f1a1188d" Sep 30 14:17:32 crc kubenswrapper[4840]: I0930 14:17:32.766738 4840 scope.go:117] "RemoveContainer" containerID="74f7b95cbd5755a9bf98cf8484373a4d51bf459301fc9332cd6c18d346695043" Sep 30 14:17:34 crc kubenswrapper[4840]: I0930 14:17:34.016101 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Sep 30 14:17:34 crc kubenswrapper[4840]: I0930 14:17:34.914759 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:17:42 crc kubenswrapper[4840]: I0930 14:17:42.797977 4840 generic.go:334] "Generic (PLEG): container finished" podID="c1073e6a-94e1-4fad-a116-ff4e145a0d22" containerID="1d44e53a0dc10a4da37ddedd8788029adf23a290f62e570cbc12812dad290363" exitCode=0 Sep 30 14:17:42 crc kubenswrapper[4840]: I0930 14:17:42.798122 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q5t88" event={"ID":"c1073e6a-94e1-4fad-a116-ff4e145a0d22","Type":"ContainerDied","Data":"1d44e53a0dc10a4da37ddedd8788029adf23a290f62e570cbc12812dad290363"} Sep 30 14:17:44 crc kubenswrapper[4840]: I0930 14:17:44.251170 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q5t88" Sep 30 14:17:44 crc kubenswrapper[4840]: I0930 14:17:44.334260 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1073e6a-94e1-4fad-a116-ff4e145a0d22-repo-setup-combined-ca-bundle\") pod \"c1073e6a-94e1-4fad-a116-ff4e145a0d22\" (UID: \"c1073e6a-94e1-4fad-a116-ff4e145a0d22\") " Sep 30 14:17:44 crc kubenswrapper[4840]: I0930 14:17:44.334323 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c1073e6a-94e1-4fad-a116-ff4e145a0d22-ssh-key\") pod \"c1073e6a-94e1-4fad-a116-ff4e145a0d22\" (UID: \"c1073e6a-94e1-4fad-a116-ff4e145a0d22\") " Sep 30 14:17:44 crc kubenswrapper[4840]: I0930 14:17:44.334348 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c1073e6a-94e1-4fad-a116-ff4e145a0d22-inventory\") pod \"c1073e6a-94e1-4fad-a116-ff4e145a0d22\" (UID: \"c1073e6a-94e1-4fad-a116-ff4e145a0d22\") " Sep 30 14:17:44 crc kubenswrapper[4840]: I0930 14:17:44.334421 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rj4c\" (UniqueName: \"kubernetes.io/projected/c1073e6a-94e1-4fad-a116-ff4e145a0d22-kube-api-access-8rj4c\") pod \"c1073e6a-94e1-4fad-a116-ff4e145a0d22\" (UID: \"c1073e6a-94e1-4fad-a116-ff4e145a0d22\") " Sep 30 14:17:44 crc kubenswrapper[4840]: I0930 14:17:44.341918 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1073e6a-94e1-4fad-a116-ff4e145a0d22-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "c1073e6a-94e1-4fad-a116-ff4e145a0d22" (UID: "c1073e6a-94e1-4fad-a116-ff4e145a0d22"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:17:44 crc kubenswrapper[4840]: I0930 14:17:44.342826 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1073e6a-94e1-4fad-a116-ff4e145a0d22-kube-api-access-8rj4c" (OuterVolumeSpecName: "kube-api-access-8rj4c") pod "c1073e6a-94e1-4fad-a116-ff4e145a0d22" (UID: "c1073e6a-94e1-4fad-a116-ff4e145a0d22"). InnerVolumeSpecName "kube-api-access-8rj4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:17:44 crc kubenswrapper[4840]: E0930 14:17:44.367170 4840 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1073e6a-94e1-4fad-a116-ff4e145a0d22-inventory podName:c1073e6a-94e1-4fad-a116-ff4e145a0d22 nodeName:}" failed. No retries permitted until 2025-09-30 14:17:44.867133762 +0000 UTC m=+1293.496220185 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "inventory" (UniqueName: "kubernetes.io/secret/c1073e6a-94e1-4fad-a116-ff4e145a0d22-inventory") pod "c1073e6a-94e1-4fad-a116-ff4e145a0d22" (UID: "c1073e6a-94e1-4fad-a116-ff4e145a0d22") : error deleting /var/lib/kubelet/pods/c1073e6a-94e1-4fad-a116-ff4e145a0d22/volume-subpaths: remove /var/lib/kubelet/pods/c1073e6a-94e1-4fad-a116-ff4e145a0d22/volume-subpaths: no such file or directory Sep 30 14:17:44 crc kubenswrapper[4840]: I0930 14:17:44.369764 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1073e6a-94e1-4fad-a116-ff4e145a0d22-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c1073e6a-94e1-4fad-a116-ff4e145a0d22" (UID: "c1073e6a-94e1-4fad-a116-ff4e145a0d22"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:17:44 crc kubenswrapper[4840]: I0930 14:17:44.435798 4840 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1073e6a-94e1-4fad-a116-ff4e145a0d22-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:17:44 crc kubenswrapper[4840]: I0930 14:17:44.435838 4840 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c1073e6a-94e1-4fad-a116-ff4e145a0d22-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:17:44 crc kubenswrapper[4840]: I0930 14:17:44.435851 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rj4c\" (UniqueName: \"kubernetes.io/projected/c1073e6a-94e1-4fad-a116-ff4e145a0d22-kube-api-access-8rj4c\") on node \"crc\" DevicePath \"\"" Sep 30 14:17:44 crc kubenswrapper[4840]: I0930 14:17:44.828523 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q5t88" event={"ID":"c1073e6a-94e1-4fad-a116-ff4e145a0d22","Type":"ContainerDied","Data":"6eb69b7b7435ecee029914f1292c73de123d03c9dfb482ba52fa14887e09fec5"} Sep 30 14:17:44 crc kubenswrapper[4840]: I0930 14:17:44.828581 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6eb69b7b7435ecee029914f1292c73de123d03c9dfb482ba52fa14887e09fec5" Sep 30 14:17:44 crc kubenswrapper[4840]: I0930 14:17:44.828612 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q5t88" Sep 30 14:17:44 crc kubenswrapper[4840]: I0930 14:17:44.918281 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-gtxqx"] Sep 30 14:17:44 crc kubenswrapper[4840]: E0930 14:17:44.918900 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1073e6a-94e1-4fad-a116-ff4e145a0d22" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Sep 30 14:17:44 crc kubenswrapper[4840]: I0930 14:17:44.918928 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1073e6a-94e1-4fad-a116-ff4e145a0d22" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Sep 30 14:17:44 crc kubenswrapper[4840]: I0930 14:17:44.919236 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1073e6a-94e1-4fad-a116-ff4e145a0d22" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Sep 30 14:17:44 crc kubenswrapper[4840]: I0930 14:17:44.920221 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gtxqx" Sep 30 14:17:44 crc kubenswrapper[4840]: I0930 14:17:44.932338 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-gtxqx"] Sep 30 14:17:44 crc kubenswrapper[4840]: I0930 14:17:44.947105 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c1073e6a-94e1-4fad-a116-ff4e145a0d22-inventory\") pod \"c1073e6a-94e1-4fad-a116-ff4e145a0d22\" (UID: \"c1073e6a-94e1-4fad-a116-ff4e145a0d22\") " Sep 30 14:17:44 crc kubenswrapper[4840]: I0930 14:17:44.947714 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mt8t\" (UniqueName: \"kubernetes.io/projected/84f24beb-97af-44e4-b529-41138e5414ad-kube-api-access-2mt8t\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-gtxqx\" (UID: \"84f24beb-97af-44e4-b529-41138e5414ad\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gtxqx" Sep 30 14:17:44 crc kubenswrapper[4840]: I0930 14:17:44.947823 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/84f24beb-97af-44e4-b529-41138e5414ad-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-gtxqx\" (UID: \"84f24beb-97af-44e4-b529-41138e5414ad\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gtxqx" Sep 30 14:17:44 crc kubenswrapper[4840]: I0930 14:17:44.948187 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/84f24beb-97af-44e4-b529-41138e5414ad-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-gtxqx\" (UID: \"84f24beb-97af-44e4-b529-41138e5414ad\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gtxqx" Sep 30 14:17:44 crc kubenswrapper[4840]: I0930 14:17:44.952293 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1073e6a-94e1-4fad-a116-ff4e145a0d22-inventory" (OuterVolumeSpecName: "inventory") pod "c1073e6a-94e1-4fad-a116-ff4e145a0d22" (UID: "c1073e6a-94e1-4fad-a116-ff4e145a0d22"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:17:45 crc kubenswrapper[4840]: I0930 14:17:45.048993 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/84f24beb-97af-44e4-b529-41138e5414ad-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-gtxqx\" (UID: \"84f24beb-97af-44e4-b529-41138e5414ad\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gtxqx" Sep 30 14:17:45 crc kubenswrapper[4840]: I0930 14:17:45.049067 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mt8t\" (UniqueName: \"kubernetes.io/projected/84f24beb-97af-44e4-b529-41138e5414ad-kube-api-access-2mt8t\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-gtxqx\" (UID: \"84f24beb-97af-44e4-b529-41138e5414ad\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gtxqx" Sep 30 14:17:45 crc kubenswrapper[4840]: I0930 14:17:45.049151 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/84f24beb-97af-44e4-b529-41138e5414ad-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-gtxqx\" (UID: \"84f24beb-97af-44e4-b529-41138e5414ad\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gtxqx" Sep 30 14:17:45 crc kubenswrapper[4840]: I0930 14:17:45.049277 4840 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c1073e6a-94e1-4fad-a116-ff4e145a0d22-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 14:17:45 crc kubenswrapper[4840]: I0930 14:17:45.053035 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/84f24beb-97af-44e4-b529-41138e5414ad-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-gtxqx\" (UID: \"84f24beb-97af-44e4-b529-41138e5414ad\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gtxqx" Sep 30 14:17:45 crc kubenswrapper[4840]: I0930 14:17:45.053039 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/84f24beb-97af-44e4-b529-41138e5414ad-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-gtxqx\" (UID: \"84f24beb-97af-44e4-b529-41138e5414ad\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gtxqx" Sep 30 14:17:45 crc kubenswrapper[4840]: I0930 14:17:45.066347 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mt8t\" (UniqueName: \"kubernetes.io/projected/84f24beb-97af-44e4-b529-41138e5414ad-kube-api-access-2mt8t\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-gtxqx\" (UID: \"84f24beb-97af-44e4-b529-41138e5414ad\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gtxqx" Sep 30 14:17:45 crc kubenswrapper[4840]: I0930 14:17:45.241610 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gtxqx" Sep 30 14:17:45 crc kubenswrapper[4840]: I0930 14:17:45.796155 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-gtxqx"] Sep 30 14:17:45 crc kubenswrapper[4840]: W0930 14:17:45.802952 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod84f24beb_97af_44e4_b529_41138e5414ad.slice/crio-e1ea0e36722668a81b41b43cb33cbbb169ddfb9f1be7cd7a5310e2c44423ea2b WatchSource:0}: Error finding container e1ea0e36722668a81b41b43cb33cbbb169ddfb9f1be7cd7a5310e2c44423ea2b: Status 404 returned error can't find the container with id e1ea0e36722668a81b41b43cb33cbbb169ddfb9f1be7cd7a5310e2c44423ea2b Sep 30 14:17:45 crc kubenswrapper[4840]: I0930 14:17:45.842838 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gtxqx" event={"ID":"84f24beb-97af-44e4-b529-41138e5414ad","Type":"ContainerStarted","Data":"e1ea0e36722668a81b41b43cb33cbbb169ddfb9f1be7cd7a5310e2c44423ea2b"} Sep 30 14:17:46 crc kubenswrapper[4840]: I0930 14:17:46.858926 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gtxqx" event={"ID":"84f24beb-97af-44e4-b529-41138e5414ad","Type":"ContainerStarted","Data":"490857b72729d820be33f052c8d09a53301b076c7b6d60b7816fd3ae6962124e"} Sep 30 14:17:49 crc kubenswrapper[4840]: I0930 14:17:49.891075 4840 generic.go:334] "Generic (PLEG): container finished" podID="84f24beb-97af-44e4-b529-41138e5414ad" containerID="490857b72729d820be33f052c8d09a53301b076c7b6d60b7816fd3ae6962124e" exitCode=0 Sep 30 14:17:49 crc kubenswrapper[4840]: I0930 14:17:49.891171 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gtxqx" event={"ID":"84f24beb-97af-44e4-b529-41138e5414ad","Type":"ContainerDied","Data":"490857b72729d820be33f052c8d09a53301b076c7b6d60b7816fd3ae6962124e"} Sep 30 14:17:51 crc kubenswrapper[4840]: I0930 14:17:51.310680 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gtxqx" Sep 30 14:17:51 crc kubenswrapper[4840]: I0930 14:17:51.479799 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/84f24beb-97af-44e4-b529-41138e5414ad-ssh-key\") pod \"84f24beb-97af-44e4-b529-41138e5414ad\" (UID: \"84f24beb-97af-44e4-b529-41138e5414ad\") " Sep 30 14:17:51 crc kubenswrapper[4840]: I0930 14:17:51.479904 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mt8t\" (UniqueName: \"kubernetes.io/projected/84f24beb-97af-44e4-b529-41138e5414ad-kube-api-access-2mt8t\") pod \"84f24beb-97af-44e4-b529-41138e5414ad\" (UID: \"84f24beb-97af-44e4-b529-41138e5414ad\") " Sep 30 14:17:51 crc kubenswrapper[4840]: I0930 14:17:51.480910 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/84f24beb-97af-44e4-b529-41138e5414ad-inventory\") pod \"84f24beb-97af-44e4-b529-41138e5414ad\" (UID: \"84f24beb-97af-44e4-b529-41138e5414ad\") " Sep 30 14:17:51 crc kubenswrapper[4840]: I0930 14:17:51.487046 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84f24beb-97af-44e4-b529-41138e5414ad-kube-api-access-2mt8t" (OuterVolumeSpecName: "kube-api-access-2mt8t") pod "84f24beb-97af-44e4-b529-41138e5414ad" (UID: "84f24beb-97af-44e4-b529-41138e5414ad"). InnerVolumeSpecName "kube-api-access-2mt8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:17:51 crc kubenswrapper[4840]: I0930 14:17:51.515124 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84f24beb-97af-44e4-b529-41138e5414ad-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "84f24beb-97af-44e4-b529-41138e5414ad" (UID: "84f24beb-97af-44e4-b529-41138e5414ad"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:17:51 crc kubenswrapper[4840]: I0930 14:17:51.515674 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84f24beb-97af-44e4-b529-41138e5414ad-inventory" (OuterVolumeSpecName: "inventory") pod "84f24beb-97af-44e4-b529-41138e5414ad" (UID: "84f24beb-97af-44e4-b529-41138e5414ad"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:17:51 crc kubenswrapper[4840]: I0930 14:17:51.583518 4840 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/84f24beb-97af-44e4-b529-41138e5414ad-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:17:51 crc kubenswrapper[4840]: I0930 14:17:51.583582 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2mt8t\" (UniqueName: \"kubernetes.io/projected/84f24beb-97af-44e4-b529-41138e5414ad-kube-api-access-2mt8t\") on node \"crc\" DevicePath \"\"" Sep 30 14:17:51 crc kubenswrapper[4840]: I0930 14:17:51.583596 4840 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/84f24beb-97af-44e4-b529-41138e5414ad-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 14:17:51 crc kubenswrapper[4840]: I0930 14:17:51.914134 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gtxqx" event={"ID":"84f24beb-97af-44e4-b529-41138e5414ad","Type":"ContainerDied","Data":"e1ea0e36722668a81b41b43cb33cbbb169ddfb9f1be7cd7a5310e2c44423ea2b"} Sep 30 14:17:51 crc kubenswrapper[4840]: I0930 14:17:51.914175 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1ea0e36722668a81b41b43cb33cbbb169ddfb9f1be7cd7a5310e2c44423ea2b" Sep 30 14:17:51 crc kubenswrapper[4840]: I0930 14:17:51.914183 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gtxqx" Sep 30 14:17:51 crc kubenswrapper[4840]: I0930 14:17:51.982718 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6ntwv"] Sep 30 14:17:51 crc kubenswrapper[4840]: E0930 14:17:51.983331 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84f24beb-97af-44e4-b529-41138e5414ad" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Sep 30 14:17:51 crc kubenswrapper[4840]: I0930 14:17:51.983365 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="84f24beb-97af-44e4-b529-41138e5414ad" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Sep 30 14:17:51 crc kubenswrapper[4840]: I0930 14:17:51.983716 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="84f24beb-97af-44e4-b529-41138e5414ad" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Sep 30 14:17:51 crc kubenswrapper[4840]: I0930 14:17:51.984591 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6ntwv" Sep 30 14:17:51 crc kubenswrapper[4840]: I0930 14:17:51.986378 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 14:17:51 crc kubenswrapper[4840]: I0930 14:17:51.986757 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 14:17:51 crc kubenswrapper[4840]: I0930 14:17:51.986937 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 14:17:51 crc kubenswrapper[4840]: I0930 14:17:51.989728 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9zm9z" Sep 30 14:17:51 crc kubenswrapper[4840]: I0930 14:17:51.993127 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6ntwv"] Sep 30 14:17:52 crc kubenswrapper[4840]: I0930 14:17:52.093382 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/558c75bb-7947-423d-b3aa-1b2ab4ba2d1d-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-6ntwv\" (UID: \"558c75bb-7947-423d-b3aa-1b2ab4ba2d1d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6ntwv" Sep 30 14:17:52 crc kubenswrapper[4840]: I0930 14:17:52.093438 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbpqb\" (UniqueName: \"kubernetes.io/projected/558c75bb-7947-423d-b3aa-1b2ab4ba2d1d-kube-api-access-mbpqb\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-6ntwv\" (UID: \"558c75bb-7947-423d-b3aa-1b2ab4ba2d1d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6ntwv" Sep 30 14:17:52 crc kubenswrapper[4840]: I0930 14:17:52.093495 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/558c75bb-7947-423d-b3aa-1b2ab4ba2d1d-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-6ntwv\" (UID: \"558c75bb-7947-423d-b3aa-1b2ab4ba2d1d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6ntwv" Sep 30 14:17:52 crc kubenswrapper[4840]: I0930 14:17:52.093575 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/558c75bb-7947-423d-b3aa-1b2ab4ba2d1d-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-6ntwv\" (UID: \"558c75bb-7947-423d-b3aa-1b2ab4ba2d1d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6ntwv" Sep 30 14:17:52 crc kubenswrapper[4840]: I0930 14:17:52.195160 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/558c75bb-7947-423d-b3aa-1b2ab4ba2d1d-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-6ntwv\" (UID: \"558c75bb-7947-423d-b3aa-1b2ab4ba2d1d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6ntwv" Sep 30 14:17:52 crc kubenswrapper[4840]: I0930 14:17:52.195318 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/558c75bb-7947-423d-b3aa-1b2ab4ba2d1d-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-6ntwv\" (UID: \"558c75bb-7947-423d-b3aa-1b2ab4ba2d1d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6ntwv" Sep 30 14:17:52 crc kubenswrapper[4840]: I0930 14:17:52.195510 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/558c75bb-7947-423d-b3aa-1b2ab4ba2d1d-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-6ntwv\" (UID: \"558c75bb-7947-423d-b3aa-1b2ab4ba2d1d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6ntwv" Sep 30 14:17:52 crc kubenswrapper[4840]: I0930 14:17:52.195592 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbpqb\" (UniqueName: \"kubernetes.io/projected/558c75bb-7947-423d-b3aa-1b2ab4ba2d1d-kube-api-access-mbpqb\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-6ntwv\" (UID: \"558c75bb-7947-423d-b3aa-1b2ab4ba2d1d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6ntwv" Sep 30 14:17:52 crc kubenswrapper[4840]: I0930 14:17:52.201761 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/558c75bb-7947-423d-b3aa-1b2ab4ba2d1d-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-6ntwv\" (UID: \"558c75bb-7947-423d-b3aa-1b2ab4ba2d1d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6ntwv" Sep 30 14:17:52 crc kubenswrapper[4840]: I0930 14:17:52.201777 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/558c75bb-7947-423d-b3aa-1b2ab4ba2d1d-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-6ntwv\" (UID: \"558c75bb-7947-423d-b3aa-1b2ab4ba2d1d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6ntwv" Sep 30 14:17:52 crc kubenswrapper[4840]: I0930 14:17:52.202508 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/558c75bb-7947-423d-b3aa-1b2ab4ba2d1d-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-6ntwv\" (UID: \"558c75bb-7947-423d-b3aa-1b2ab4ba2d1d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6ntwv" Sep 30 14:17:52 crc kubenswrapper[4840]: I0930 14:17:52.217625 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbpqb\" (UniqueName: \"kubernetes.io/projected/558c75bb-7947-423d-b3aa-1b2ab4ba2d1d-kube-api-access-mbpqb\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-6ntwv\" (UID: \"558c75bb-7947-423d-b3aa-1b2ab4ba2d1d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6ntwv" Sep 30 14:17:52 crc kubenswrapper[4840]: I0930 14:17:52.302475 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6ntwv" Sep 30 14:17:52 crc kubenswrapper[4840]: I0930 14:17:52.820911 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6ntwv"] Sep 30 14:17:52 crc kubenswrapper[4840]: I0930 14:17:52.927296 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6ntwv" event={"ID":"558c75bb-7947-423d-b3aa-1b2ab4ba2d1d","Type":"ContainerStarted","Data":"0df5995a0b3eb6a76478cd96d84419785237ae2ff95b871f1ffdc1e510433c10"} Sep 30 14:17:53 crc kubenswrapper[4840]: I0930 14:17:53.940839 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6ntwv" event={"ID":"558c75bb-7947-423d-b3aa-1b2ab4ba2d1d","Type":"ContainerStarted","Data":"f5a2f1cea1343ab9cbdd93d48463fd9a34407ef0712526e2b7a909c4919b614e"} Sep 30 14:17:53 crc kubenswrapper[4840]: I0930 14:17:53.963146 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6ntwv" podStartSLOduration=2.185892695 podStartE2EDuration="2.963125394s" podCreationTimestamp="2025-09-30 14:17:51 +0000 UTC" firstStartedPulling="2025-09-30 14:17:52.827757524 +0000 UTC m=+1301.456843947" lastFinishedPulling="2025-09-30 14:17:53.604990203 +0000 UTC m=+1302.234076646" observedRunningTime="2025-09-30 14:17:53.955788946 +0000 UTC m=+1302.584875399" watchObservedRunningTime="2025-09-30 14:17:53.963125394 +0000 UTC m=+1302.592211817" Sep 30 14:18:32 crc kubenswrapper[4840]: I0930 14:18:32.906317 4840 scope.go:117] "RemoveContainer" containerID="3e4d847f6209271b5d7feb61da4bc7a0a0745742755f37fd279116bb4f538388" Sep 30 14:18:32 crc kubenswrapper[4840]: I0930 14:18:32.927753 4840 scope.go:117] "RemoveContainer" containerID="cc38a3c8c3fdff136f33150d801d31f8f5cf2972029015e0b6998146246f4c35" Sep 30 14:18:32 crc kubenswrapper[4840]: I0930 14:18:32.958101 4840 scope.go:117] "RemoveContainer" containerID="4f71635a7189cd95395acdf5d262684f681fe8e3cb9c262b0c1d2ac9633e1d4a" Sep 30 14:18:33 crc kubenswrapper[4840]: I0930 14:18:33.015879 4840 scope.go:117] "RemoveContainer" containerID="9bbf388fd07ba823ead856edc82d164211722304b951e16bdab0ea953989ade1" Sep 30 14:18:33 crc kubenswrapper[4840]: I0930 14:18:33.062675 4840 scope.go:117] "RemoveContainer" containerID="81d2e9ea2e06c3dd73796140bb1862f4d42eeb2f12f035a1619c6007f2483bb7" Sep 30 14:19:31 crc kubenswrapper[4840]: I0930 14:19:31.600356 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-b28jj"] Sep 30 14:19:31 crc kubenswrapper[4840]: I0930 14:19:31.603013 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b28jj" Sep 30 14:19:31 crc kubenswrapper[4840]: I0930 14:19:31.611709 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b28jj"] Sep 30 14:19:31 crc kubenswrapper[4840]: I0930 14:19:31.786743 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2db8b813-bd3a-428b-8cfa-be0422c204e5-catalog-content\") pod \"certified-operators-b28jj\" (UID: \"2db8b813-bd3a-428b-8cfa-be0422c204e5\") " pod="openshift-marketplace/certified-operators-b28jj" Sep 30 14:19:31 crc kubenswrapper[4840]: I0930 14:19:31.786888 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2db8b813-bd3a-428b-8cfa-be0422c204e5-utilities\") pod \"certified-operators-b28jj\" (UID: \"2db8b813-bd3a-428b-8cfa-be0422c204e5\") " pod="openshift-marketplace/certified-operators-b28jj" Sep 30 14:19:31 crc kubenswrapper[4840]: I0930 14:19:31.786939 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxgc6\" (UniqueName: \"kubernetes.io/projected/2db8b813-bd3a-428b-8cfa-be0422c204e5-kube-api-access-zxgc6\") pod \"certified-operators-b28jj\" (UID: \"2db8b813-bd3a-428b-8cfa-be0422c204e5\") " pod="openshift-marketplace/certified-operators-b28jj" Sep 30 14:19:31 crc kubenswrapper[4840]: I0930 14:19:31.889082 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2db8b813-bd3a-428b-8cfa-be0422c204e5-catalog-content\") pod \"certified-operators-b28jj\" (UID: \"2db8b813-bd3a-428b-8cfa-be0422c204e5\") " pod="openshift-marketplace/certified-operators-b28jj" Sep 30 14:19:31 crc kubenswrapper[4840]: I0930 14:19:31.889216 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2db8b813-bd3a-428b-8cfa-be0422c204e5-utilities\") pod \"certified-operators-b28jj\" (UID: \"2db8b813-bd3a-428b-8cfa-be0422c204e5\") " pod="openshift-marketplace/certified-operators-b28jj" Sep 30 14:19:31 crc kubenswrapper[4840]: I0930 14:19:31.889278 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxgc6\" (UniqueName: \"kubernetes.io/projected/2db8b813-bd3a-428b-8cfa-be0422c204e5-kube-api-access-zxgc6\") pod \"certified-operators-b28jj\" (UID: \"2db8b813-bd3a-428b-8cfa-be0422c204e5\") " pod="openshift-marketplace/certified-operators-b28jj" Sep 30 14:19:31 crc kubenswrapper[4840]: I0930 14:19:31.889618 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2db8b813-bd3a-428b-8cfa-be0422c204e5-catalog-content\") pod \"certified-operators-b28jj\" (UID: \"2db8b813-bd3a-428b-8cfa-be0422c204e5\") " pod="openshift-marketplace/certified-operators-b28jj" Sep 30 14:19:31 crc kubenswrapper[4840]: I0930 14:19:31.889723 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2db8b813-bd3a-428b-8cfa-be0422c204e5-utilities\") pod \"certified-operators-b28jj\" (UID: \"2db8b813-bd3a-428b-8cfa-be0422c204e5\") " pod="openshift-marketplace/certified-operators-b28jj" Sep 30 14:19:31 crc kubenswrapper[4840]: I0930 14:19:31.909446 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxgc6\" (UniqueName: \"kubernetes.io/projected/2db8b813-bd3a-428b-8cfa-be0422c204e5-kube-api-access-zxgc6\") pod \"certified-operators-b28jj\" (UID: \"2db8b813-bd3a-428b-8cfa-be0422c204e5\") " pod="openshift-marketplace/certified-operators-b28jj" Sep 30 14:19:31 crc kubenswrapper[4840]: I0930 14:19:31.921978 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b28jj" Sep 30 14:19:32 crc kubenswrapper[4840]: I0930 14:19:32.412097 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b28jj"] Sep 30 14:19:32 crc kubenswrapper[4840]: W0930 14:19:32.424836 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2db8b813_bd3a_428b_8cfa_be0422c204e5.slice/crio-6149893c0563921311853d552852cc50202e92f405b09c01ae99eea79eb070e0 WatchSource:0}: Error finding container 6149893c0563921311853d552852cc50202e92f405b09c01ae99eea79eb070e0: Status 404 returned error can't find the container with id 6149893c0563921311853d552852cc50202e92f405b09c01ae99eea79eb070e0 Sep 30 14:19:32 crc kubenswrapper[4840]: I0930 14:19:32.901335 4840 generic.go:334] "Generic (PLEG): container finished" podID="2db8b813-bd3a-428b-8cfa-be0422c204e5" containerID="78f6185815f968edc83336753bc41b2809a0ed538db934f28be058a56872d951" exitCode=0 Sep 30 14:19:32 crc kubenswrapper[4840]: I0930 14:19:32.901383 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b28jj" event={"ID":"2db8b813-bd3a-428b-8cfa-be0422c204e5","Type":"ContainerDied","Data":"78f6185815f968edc83336753bc41b2809a0ed538db934f28be058a56872d951"} Sep 30 14:19:32 crc kubenswrapper[4840]: I0930 14:19:32.901434 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b28jj" event={"ID":"2db8b813-bd3a-428b-8cfa-be0422c204e5","Type":"ContainerStarted","Data":"6149893c0563921311853d552852cc50202e92f405b09c01ae99eea79eb070e0"} Sep 30 14:19:33 crc kubenswrapper[4840]: I0930 14:19:33.164021 4840 scope.go:117] "RemoveContainer" containerID="9f12e56d9a170038e99c98ffb6f13331191d2fdd43eecac910f50328d12487f7" Sep 30 14:19:34 crc kubenswrapper[4840]: I0930 14:19:34.923500 4840 generic.go:334] "Generic (PLEG): container finished" podID="2db8b813-bd3a-428b-8cfa-be0422c204e5" containerID="ae6bcf437adfc13773608b5091c05274dc480933577b405ed2de91380276d761" exitCode=0 Sep 30 14:19:34 crc kubenswrapper[4840]: I0930 14:19:34.923608 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b28jj" event={"ID":"2db8b813-bd3a-428b-8cfa-be0422c204e5","Type":"ContainerDied","Data":"ae6bcf437adfc13773608b5091c05274dc480933577b405ed2de91380276d761"} Sep 30 14:19:35 crc kubenswrapper[4840]: I0930 14:19:35.937891 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b28jj" event={"ID":"2db8b813-bd3a-428b-8cfa-be0422c204e5","Type":"ContainerStarted","Data":"56009700ffa6341431994c82915ecd612dfa17cd10e383753d2a1e181dca807e"} Sep 30 14:19:35 crc kubenswrapper[4840]: I0930 14:19:35.963961 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-b28jj" podStartSLOduration=2.530061465 podStartE2EDuration="4.963942701s" podCreationTimestamp="2025-09-30 14:19:31 +0000 UTC" firstStartedPulling="2025-09-30 14:19:32.903068227 +0000 UTC m=+1401.532154650" lastFinishedPulling="2025-09-30 14:19:35.336949463 +0000 UTC m=+1403.966035886" observedRunningTime="2025-09-30 14:19:35.956917049 +0000 UTC m=+1404.586003492" watchObservedRunningTime="2025-09-30 14:19:35.963942701 +0000 UTC m=+1404.593029124" Sep 30 14:19:41 crc kubenswrapper[4840]: I0930 14:19:41.922617 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-b28jj" Sep 30 14:19:41 crc kubenswrapper[4840]: I0930 14:19:41.923103 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-b28jj" Sep 30 14:19:41 crc kubenswrapper[4840]: I0930 14:19:41.975925 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-b28jj" Sep 30 14:19:42 crc kubenswrapper[4840]: I0930 14:19:42.052151 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-b28jj" Sep 30 14:19:42 crc kubenswrapper[4840]: I0930 14:19:42.216797 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b28jj"] Sep 30 14:19:44 crc kubenswrapper[4840]: I0930 14:19:44.009793 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-b28jj" podUID="2db8b813-bd3a-428b-8cfa-be0422c204e5" containerName="registry-server" containerID="cri-o://56009700ffa6341431994c82915ecd612dfa17cd10e383753d2a1e181dca807e" gracePeriod=2 Sep 30 14:19:44 crc kubenswrapper[4840]: I0930 14:19:44.455612 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b28jj" Sep 30 14:19:44 crc kubenswrapper[4840]: I0930 14:19:44.528928 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxgc6\" (UniqueName: \"kubernetes.io/projected/2db8b813-bd3a-428b-8cfa-be0422c204e5-kube-api-access-zxgc6\") pod \"2db8b813-bd3a-428b-8cfa-be0422c204e5\" (UID: \"2db8b813-bd3a-428b-8cfa-be0422c204e5\") " Sep 30 14:19:44 crc kubenswrapper[4840]: I0930 14:19:44.529226 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2db8b813-bd3a-428b-8cfa-be0422c204e5-utilities\") pod \"2db8b813-bd3a-428b-8cfa-be0422c204e5\" (UID: \"2db8b813-bd3a-428b-8cfa-be0422c204e5\") " Sep 30 14:19:44 crc kubenswrapper[4840]: I0930 14:19:44.529281 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2db8b813-bd3a-428b-8cfa-be0422c204e5-catalog-content\") pod \"2db8b813-bd3a-428b-8cfa-be0422c204e5\" (UID: \"2db8b813-bd3a-428b-8cfa-be0422c204e5\") " Sep 30 14:19:44 crc kubenswrapper[4840]: I0930 14:19:44.530253 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2db8b813-bd3a-428b-8cfa-be0422c204e5-utilities" (OuterVolumeSpecName: "utilities") pod "2db8b813-bd3a-428b-8cfa-be0422c204e5" (UID: "2db8b813-bd3a-428b-8cfa-be0422c204e5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:19:44 crc kubenswrapper[4840]: I0930 14:19:44.534202 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2db8b813-bd3a-428b-8cfa-be0422c204e5-kube-api-access-zxgc6" (OuterVolumeSpecName: "kube-api-access-zxgc6") pod "2db8b813-bd3a-428b-8cfa-be0422c204e5" (UID: "2db8b813-bd3a-428b-8cfa-be0422c204e5"). InnerVolumeSpecName "kube-api-access-zxgc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:44 crc kubenswrapper[4840]: I0930 14:19:44.580919 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2db8b813-bd3a-428b-8cfa-be0422c204e5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2db8b813-bd3a-428b-8cfa-be0422c204e5" (UID: "2db8b813-bd3a-428b-8cfa-be0422c204e5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:19:44 crc kubenswrapper[4840]: I0930 14:19:44.631284 4840 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2db8b813-bd3a-428b-8cfa-be0422c204e5-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:44 crc kubenswrapper[4840]: I0930 14:19:44.631321 4840 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2db8b813-bd3a-428b-8cfa-be0422c204e5-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:44 crc kubenswrapper[4840]: I0930 14:19:44.631331 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxgc6\" (UniqueName: \"kubernetes.io/projected/2db8b813-bd3a-428b-8cfa-be0422c204e5-kube-api-access-zxgc6\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:45 crc kubenswrapper[4840]: I0930 14:19:45.025192 4840 generic.go:334] "Generic (PLEG): container finished" podID="2db8b813-bd3a-428b-8cfa-be0422c204e5" containerID="56009700ffa6341431994c82915ecd612dfa17cd10e383753d2a1e181dca807e" exitCode=0 Sep 30 14:19:45 crc kubenswrapper[4840]: I0930 14:19:45.025310 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b28jj" Sep 30 14:19:45 crc kubenswrapper[4840]: I0930 14:19:45.025329 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b28jj" event={"ID":"2db8b813-bd3a-428b-8cfa-be0422c204e5","Type":"ContainerDied","Data":"56009700ffa6341431994c82915ecd612dfa17cd10e383753d2a1e181dca807e"} Sep 30 14:19:45 crc kubenswrapper[4840]: I0930 14:19:45.025756 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b28jj" event={"ID":"2db8b813-bd3a-428b-8cfa-be0422c204e5","Type":"ContainerDied","Data":"6149893c0563921311853d552852cc50202e92f405b09c01ae99eea79eb070e0"} Sep 30 14:19:45 crc kubenswrapper[4840]: I0930 14:19:45.025787 4840 scope.go:117] "RemoveContainer" containerID="56009700ffa6341431994c82915ecd612dfa17cd10e383753d2a1e181dca807e" Sep 30 14:19:45 crc kubenswrapper[4840]: I0930 14:19:45.048996 4840 scope.go:117] "RemoveContainer" containerID="ae6bcf437adfc13773608b5091c05274dc480933577b405ed2de91380276d761" Sep 30 14:19:45 crc kubenswrapper[4840]: I0930 14:19:45.065004 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b28jj"] Sep 30 14:19:45 crc kubenswrapper[4840]: I0930 14:19:45.073903 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-b28jj"] Sep 30 14:19:45 crc kubenswrapper[4840]: I0930 14:19:45.089675 4840 scope.go:117] "RemoveContainer" containerID="78f6185815f968edc83336753bc41b2809a0ed538db934f28be058a56872d951" Sep 30 14:19:45 crc kubenswrapper[4840]: I0930 14:19:45.118218 4840 scope.go:117] "RemoveContainer" containerID="56009700ffa6341431994c82915ecd612dfa17cd10e383753d2a1e181dca807e" Sep 30 14:19:45 crc kubenswrapper[4840]: E0930 14:19:45.118795 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56009700ffa6341431994c82915ecd612dfa17cd10e383753d2a1e181dca807e\": container with ID starting with 56009700ffa6341431994c82915ecd612dfa17cd10e383753d2a1e181dca807e not found: ID does not exist" containerID="56009700ffa6341431994c82915ecd612dfa17cd10e383753d2a1e181dca807e" Sep 30 14:19:45 crc kubenswrapper[4840]: I0930 14:19:45.118843 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56009700ffa6341431994c82915ecd612dfa17cd10e383753d2a1e181dca807e"} err="failed to get container status \"56009700ffa6341431994c82915ecd612dfa17cd10e383753d2a1e181dca807e\": rpc error: code = NotFound desc = could not find container \"56009700ffa6341431994c82915ecd612dfa17cd10e383753d2a1e181dca807e\": container with ID starting with 56009700ffa6341431994c82915ecd612dfa17cd10e383753d2a1e181dca807e not found: ID does not exist" Sep 30 14:19:45 crc kubenswrapper[4840]: I0930 14:19:45.118871 4840 scope.go:117] "RemoveContainer" containerID="ae6bcf437adfc13773608b5091c05274dc480933577b405ed2de91380276d761" Sep 30 14:19:45 crc kubenswrapper[4840]: E0930 14:19:45.119155 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae6bcf437adfc13773608b5091c05274dc480933577b405ed2de91380276d761\": container with ID starting with ae6bcf437adfc13773608b5091c05274dc480933577b405ed2de91380276d761 not found: ID does not exist" containerID="ae6bcf437adfc13773608b5091c05274dc480933577b405ed2de91380276d761" Sep 30 14:19:45 crc kubenswrapper[4840]: I0930 14:19:45.119182 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae6bcf437adfc13773608b5091c05274dc480933577b405ed2de91380276d761"} err="failed to get container status \"ae6bcf437adfc13773608b5091c05274dc480933577b405ed2de91380276d761\": rpc error: code = NotFound desc = could not find container \"ae6bcf437adfc13773608b5091c05274dc480933577b405ed2de91380276d761\": container with ID starting with ae6bcf437adfc13773608b5091c05274dc480933577b405ed2de91380276d761 not found: ID does not exist" Sep 30 14:19:45 crc kubenswrapper[4840]: I0930 14:19:45.119200 4840 scope.go:117] "RemoveContainer" containerID="78f6185815f968edc83336753bc41b2809a0ed538db934f28be058a56872d951" Sep 30 14:19:45 crc kubenswrapper[4840]: E0930 14:19:45.119461 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78f6185815f968edc83336753bc41b2809a0ed538db934f28be058a56872d951\": container with ID starting with 78f6185815f968edc83336753bc41b2809a0ed538db934f28be058a56872d951 not found: ID does not exist" containerID="78f6185815f968edc83336753bc41b2809a0ed538db934f28be058a56872d951" Sep 30 14:19:45 crc kubenswrapper[4840]: I0930 14:19:45.119513 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78f6185815f968edc83336753bc41b2809a0ed538db934f28be058a56872d951"} err="failed to get container status \"78f6185815f968edc83336753bc41b2809a0ed538db934f28be058a56872d951\": rpc error: code = NotFound desc = could not find container \"78f6185815f968edc83336753bc41b2809a0ed538db934f28be058a56872d951\": container with ID starting with 78f6185815f968edc83336753bc41b2809a0ed538db934f28be058a56872d951 not found: ID does not exist" Sep 30 14:19:46 crc kubenswrapper[4840]: I0930 14:19:46.129481 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2db8b813-bd3a-428b-8cfa-be0422c204e5" path="/var/lib/kubelet/pods/2db8b813-bd3a-428b-8cfa-be0422c204e5/volumes" Sep 30 14:19:51 crc kubenswrapper[4840]: I0930 14:19:51.871729 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:19:51 crc kubenswrapper[4840]: I0930 14:19:51.872325 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:20:21 crc kubenswrapper[4840]: I0930 14:20:21.872098 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:20:21 crc kubenswrapper[4840]: I0930 14:20:21.872754 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:20:29 crc kubenswrapper[4840]: I0930 14:20:29.252933 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rd7k5"] Sep 30 14:20:29 crc kubenswrapper[4840]: E0930 14:20:29.255290 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2db8b813-bd3a-428b-8cfa-be0422c204e5" containerName="registry-server" Sep 30 14:20:29 crc kubenswrapper[4840]: I0930 14:20:29.255398 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="2db8b813-bd3a-428b-8cfa-be0422c204e5" containerName="registry-server" Sep 30 14:20:29 crc kubenswrapper[4840]: E0930 14:20:29.255497 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2db8b813-bd3a-428b-8cfa-be0422c204e5" containerName="extract-utilities" Sep 30 14:20:29 crc kubenswrapper[4840]: I0930 14:20:29.255588 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="2db8b813-bd3a-428b-8cfa-be0422c204e5" containerName="extract-utilities" Sep 30 14:20:29 crc kubenswrapper[4840]: E0930 14:20:29.256375 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2db8b813-bd3a-428b-8cfa-be0422c204e5" containerName="extract-content" Sep 30 14:20:29 crc kubenswrapper[4840]: I0930 14:20:29.256471 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="2db8b813-bd3a-428b-8cfa-be0422c204e5" containerName="extract-content" Sep 30 14:20:29 crc kubenswrapper[4840]: I0930 14:20:29.256871 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="2db8b813-bd3a-428b-8cfa-be0422c204e5" containerName="registry-server" Sep 30 14:20:29 crc kubenswrapper[4840]: I0930 14:20:29.258915 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rd7k5" Sep 30 14:20:29 crc kubenswrapper[4840]: I0930 14:20:29.267011 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rd7k5"] Sep 30 14:20:29 crc kubenswrapper[4840]: I0930 14:20:29.383878 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grr2m\" (UniqueName: \"kubernetes.io/projected/efd5ed09-c17d-4aed-970a-4b85e5b1ef2d-kube-api-access-grr2m\") pod \"community-operators-rd7k5\" (UID: \"efd5ed09-c17d-4aed-970a-4b85e5b1ef2d\") " pod="openshift-marketplace/community-operators-rd7k5" Sep 30 14:20:29 crc kubenswrapper[4840]: I0930 14:20:29.383975 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efd5ed09-c17d-4aed-970a-4b85e5b1ef2d-utilities\") pod \"community-operators-rd7k5\" (UID: \"efd5ed09-c17d-4aed-970a-4b85e5b1ef2d\") " pod="openshift-marketplace/community-operators-rd7k5" Sep 30 14:20:29 crc kubenswrapper[4840]: I0930 14:20:29.384096 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efd5ed09-c17d-4aed-970a-4b85e5b1ef2d-catalog-content\") pod \"community-operators-rd7k5\" (UID: \"efd5ed09-c17d-4aed-970a-4b85e5b1ef2d\") " pod="openshift-marketplace/community-operators-rd7k5" Sep 30 14:20:29 crc kubenswrapper[4840]: I0930 14:20:29.485714 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grr2m\" (UniqueName: \"kubernetes.io/projected/efd5ed09-c17d-4aed-970a-4b85e5b1ef2d-kube-api-access-grr2m\") pod \"community-operators-rd7k5\" (UID: \"efd5ed09-c17d-4aed-970a-4b85e5b1ef2d\") " pod="openshift-marketplace/community-operators-rd7k5" Sep 30 14:20:29 crc kubenswrapper[4840]: I0930 14:20:29.485992 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efd5ed09-c17d-4aed-970a-4b85e5b1ef2d-utilities\") pod \"community-operators-rd7k5\" (UID: \"efd5ed09-c17d-4aed-970a-4b85e5b1ef2d\") " pod="openshift-marketplace/community-operators-rd7k5" Sep 30 14:20:29 crc kubenswrapper[4840]: I0930 14:20:29.486187 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efd5ed09-c17d-4aed-970a-4b85e5b1ef2d-catalog-content\") pod \"community-operators-rd7k5\" (UID: \"efd5ed09-c17d-4aed-970a-4b85e5b1ef2d\") " pod="openshift-marketplace/community-operators-rd7k5" Sep 30 14:20:29 crc kubenswrapper[4840]: I0930 14:20:29.486478 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efd5ed09-c17d-4aed-970a-4b85e5b1ef2d-utilities\") pod \"community-operators-rd7k5\" (UID: \"efd5ed09-c17d-4aed-970a-4b85e5b1ef2d\") " pod="openshift-marketplace/community-operators-rd7k5" Sep 30 14:20:29 crc kubenswrapper[4840]: I0930 14:20:29.486521 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efd5ed09-c17d-4aed-970a-4b85e5b1ef2d-catalog-content\") pod \"community-operators-rd7k5\" (UID: \"efd5ed09-c17d-4aed-970a-4b85e5b1ef2d\") " pod="openshift-marketplace/community-operators-rd7k5" Sep 30 14:20:29 crc kubenswrapper[4840]: I0930 14:20:29.506959 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grr2m\" (UniqueName: \"kubernetes.io/projected/efd5ed09-c17d-4aed-970a-4b85e5b1ef2d-kube-api-access-grr2m\") pod \"community-operators-rd7k5\" (UID: \"efd5ed09-c17d-4aed-970a-4b85e5b1ef2d\") " pod="openshift-marketplace/community-operators-rd7k5" Sep 30 14:20:29 crc kubenswrapper[4840]: I0930 14:20:29.585149 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rd7k5" Sep 30 14:20:30 crc kubenswrapper[4840]: I0930 14:20:30.141359 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rd7k5"] Sep 30 14:20:30 crc kubenswrapper[4840]: I0930 14:20:30.458717 4840 generic.go:334] "Generic (PLEG): container finished" podID="efd5ed09-c17d-4aed-970a-4b85e5b1ef2d" containerID="b947511904f690db585eb7e5bc7cf3ca431f429126a6fa7ee42c63d6a6fcbadb" exitCode=0 Sep 30 14:20:30 crc kubenswrapper[4840]: I0930 14:20:30.458936 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rd7k5" event={"ID":"efd5ed09-c17d-4aed-970a-4b85e5b1ef2d","Type":"ContainerDied","Data":"b947511904f690db585eb7e5bc7cf3ca431f429126a6fa7ee42c63d6a6fcbadb"} Sep 30 14:20:30 crc kubenswrapper[4840]: I0930 14:20:30.460973 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rd7k5" event={"ID":"efd5ed09-c17d-4aed-970a-4b85e5b1ef2d","Type":"ContainerStarted","Data":"c20eb480878f044f2e06b1b6f4be85bdfdee553c90cc57dca4c7d52b9cef31d0"} Sep 30 14:20:31 crc kubenswrapper[4840]: I0930 14:20:31.472035 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rd7k5" event={"ID":"efd5ed09-c17d-4aed-970a-4b85e5b1ef2d","Type":"ContainerStarted","Data":"8e4c26e67ac1d5d3719742bce870aef3b90f9db3f01d9ae0b0ea32ba884f743a"} Sep 30 14:20:32 crc kubenswrapper[4840]: I0930 14:20:32.486960 4840 generic.go:334] "Generic (PLEG): container finished" podID="efd5ed09-c17d-4aed-970a-4b85e5b1ef2d" containerID="8e4c26e67ac1d5d3719742bce870aef3b90f9db3f01d9ae0b0ea32ba884f743a" exitCode=0 Sep 30 14:20:32 crc kubenswrapper[4840]: I0930 14:20:32.487002 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rd7k5" event={"ID":"efd5ed09-c17d-4aed-970a-4b85e5b1ef2d","Type":"ContainerDied","Data":"8e4c26e67ac1d5d3719742bce870aef3b90f9db3f01d9ae0b0ea32ba884f743a"} Sep 30 14:20:33 crc kubenswrapper[4840]: I0930 14:20:33.497941 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rd7k5" event={"ID":"efd5ed09-c17d-4aed-970a-4b85e5b1ef2d","Type":"ContainerStarted","Data":"2308dbee2c22a4ec51f06fdbe26bc009704eb00dc91446396f26dc6627bb3ed8"} Sep 30 14:20:33 crc kubenswrapper[4840]: I0930 14:20:33.515655 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rd7k5" podStartSLOduration=1.969784292 podStartE2EDuration="4.515632915s" podCreationTimestamp="2025-09-30 14:20:29 +0000 UTC" firstStartedPulling="2025-09-30 14:20:30.461081875 +0000 UTC m=+1459.090168298" lastFinishedPulling="2025-09-30 14:20:33.006930478 +0000 UTC m=+1461.636016921" observedRunningTime="2025-09-30 14:20:33.511111819 +0000 UTC m=+1462.140198252" watchObservedRunningTime="2025-09-30 14:20:33.515632915 +0000 UTC m=+1462.144719338" Sep 30 14:20:35 crc kubenswrapper[4840]: I0930 14:20:35.242919 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-m7mq8"] Sep 30 14:20:35 crc kubenswrapper[4840]: I0930 14:20:35.245368 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m7mq8" Sep 30 14:20:35 crc kubenswrapper[4840]: I0930 14:20:35.270765 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-m7mq8"] Sep 30 14:20:35 crc kubenswrapper[4840]: I0930 14:20:35.304235 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r899z\" (UniqueName: \"kubernetes.io/projected/5e5fc6b6-b432-4d70-8704-bbd2bd637691-kube-api-access-r899z\") pod \"redhat-operators-m7mq8\" (UID: \"5e5fc6b6-b432-4d70-8704-bbd2bd637691\") " pod="openshift-marketplace/redhat-operators-m7mq8" Sep 30 14:20:35 crc kubenswrapper[4840]: I0930 14:20:35.304292 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e5fc6b6-b432-4d70-8704-bbd2bd637691-utilities\") pod \"redhat-operators-m7mq8\" (UID: \"5e5fc6b6-b432-4d70-8704-bbd2bd637691\") " pod="openshift-marketplace/redhat-operators-m7mq8" Sep 30 14:20:35 crc kubenswrapper[4840]: I0930 14:20:35.304429 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e5fc6b6-b432-4d70-8704-bbd2bd637691-catalog-content\") pod \"redhat-operators-m7mq8\" (UID: \"5e5fc6b6-b432-4d70-8704-bbd2bd637691\") " pod="openshift-marketplace/redhat-operators-m7mq8" Sep 30 14:20:35 crc kubenswrapper[4840]: I0930 14:20:35.406100 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e5fc6b6-b432-4d70-8704-bbd2bd637691-catalog-content\") pod \"redhat-operators-m7mq8\" (UID: \"5e5fc6b6-b432-4d70-8704-bbd2bd637691\") " pod="openshift-marketplace/redhat-operators-m7mq8" Sep 30 14:20:35 crc kubenswrapper[4840]: I0930 14:20:35.406465 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r899z\" (UniqueName: \"kubernetes.io/projected/5e5fc6b6-b432-4d70-8704-bbd2bd637691-kube-api-access-r899z\") pod \"redhat-operators-m7mq8\" (UID: \"5e5fc6b6-b432-4d70-8704-bbd2bd637691\") " pod="openshift-marketplace/redhat-operators-m7mq8" Sep 30 14:20:35 crc kubenswrapper[4840]: I0930 14:20:35.406585 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e5fc6b6-b432-4d70-8704-bbd2bd637691-utilities\") pod \"redhat-operators-m7mq8\" (UID: \"5e5fc6b6-b432-4d70-8704-bbd2bd637691\") " pod="openshift-marketplace/redhat-operators-m7mq8" Sep 30 14:20:35 crc kubenswrapper[4840]: I0930 14:20:35.406825 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e5fc6b6-b432-4d70-8704-bbd2bd637691-catalog-content\") pod \"redhat-operators-m7mq8\" (UID: \"5e5fc6b6-b432-4d70-8704-bbd2bd637691\") " pod="openshift-marketplace/redhat-operators-m7mq8" Sep 30 14:20:35 crc kubenswrapper[4840]: I0930 14:20:35.406884 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e5fc6b6-b432-4d70-8704-bbd2bd637691-utilities\") pod \"redhat-operators-m7mq8\" (UID: \"5e5fc6b6-b432-4d70-8704-bbd2bd637691\") " pod="openshift-marketplace/redhat-operators-m7mq8" Sep 30 14:20:35 crc kubenswrapper[4840]: I0930 14:20:35.426255 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r899z\" (UniqueName: \"kubernetes.io/projected/5e5fc6b6-b432-4d70-8704-bbd2bd637691-kube-api-access-r899z\") pod \"redhat-operators-m7mq8\" (UID: \"5e5fc6b6-b432-4d70-8704-bbd2bd637691\") " pod="openshift-marketplace/redhat-operators-m7mq8" Sep 30 14:20:35 crc kubenswrapper[4840]: I0930 14:20:35.571669 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m7mq8" Sep 30 14:20:36 crc kubenswrapper[4840]: I0930 14:20:36.016501 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-m7mq8"] Sep 30 14:20:36 crc kubenswrapper[4840]: W0930 14:20:36.053221 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e5fc6b6_b432_4d70_8704_bbd2bd637691.slice/crio-88bed3eebd86db9b64af492f6ea3ce784bc70f31127bb30ba6dccc41bc78f68a WatchSource:0}: Error finding container 88bed3eebd86db9b64af492f6ea3ce784bc70f31127bb30ba6dccc41bc78f68a: Status 404 returned error can't find the container with id 88bed3eebd86db9b64af492f6ea3ce784bc70f31127bb30ba6dccc41bc78f68a Sep 30 14:20:36 crc kubenswrapper[4840]: I0930 14:20:36.529407 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m7mq8" event={"ID":"5e5fc6b6-b432-4d70-8704-bbd2bd637691","Type":"ContainerStarted","Data":"88bed3eebd86db9b64af492f6ea3ce784bc70f31127bb30ba6dccc41bc78f68a"} Sep 30 14:20:37 crc kubenswrapper[4840]: I0930 14:20:37.543998 4840 generic.go:334] "Generic (PLEG): container finished" podID="5e5fc6b6-b432-4d70-8704-bbd2bd637691" containerID="c8c3fd0e2a9d08ae30b12f9007523d299027aada67ddae87e9a128ad08943329" exitCode=0 Sep 30 14:20:37 crc kubenswrapper[4840]: I0930 14:20:37.544071 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m7mq8" event={"ID":"5e5fc6b6-b432-4d70-8704-bbd2bd637691","Type":"ContainerDied","Data":"c8c3fd0e2a9d08ae30b12f9007523d299027aada67ddae87e9a128ad08943329"} Sep 30 14:20:38 crc kubenswrapper[4840]: I0930 14:20:38.556020 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m7mq8" event={"ID":"5e5fc6b6-b432-4d70-8704-bbd2bd637691","Type":"ContainerStarted","Data":"f4ffdc4b3980f4c74ff4f9b818950753f5e6ae92cd9a4e3b8823bcd146774b4a"} Sep 30 14:20:39 crc kubenswrapper[4840]: I0930 14:20:39.585622 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rd7k5" Sep 30 14:20:39 crc kubenswrapper[4840]: I0930 14:20:39.586014 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rd7k5" Sep 30 14:20:39 crc kubenswrapper[4840]: I0930 14:20:39.631561 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rd7k5" Sep 30 14:20:40 crc kubenswrapper[4840]: I0930 14:20:40.637023 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rd7k5" Sep 30 14:20:41 crc kubenswrapper[4840]: I0930 14:20:41.828443 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rd7k5"] Sep 30 14:20:42 crc kubenswrapper[4840]: I0930 14:20:42.595340 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rd7k5" podUID="efd5ed09-c17d-4aed-970a-4b85e5b1ef2d" containerName="registry-server" containerID="cri-o://2308dbee2c22a4ec51f06fdbe26bc009704eb00dc91446396f26dc6627bb3ed8" gracePeriod=2 Sep 30 14:20:43 crc kubenswrapper[4840]: I0930 14:20:43.035624 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rd7k5" Sep 30 14:20:43 crc kubenswrapper[4840]: I0930 14:20:43.050882 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efd5ed09-c17d-4aed-970a-4b85e5b1ef2d-catalog-content\") pod \"efd5ed09-c17d-4aed-970a-4b85e5b1ef2d\" (UID: \"efd5ed09-c17d-4aed-970a-4b85e5b1ef2d\") " Sep 30 14:20:43 crc kubenswrapper[4840]: I0930 14:20:43.051274 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efd5ed09-c17d-4aed-970a-4b85e5b1ef2d-utilities\") pod \"efd5ed09-c17d-4aed-970a-4b85e5b1ef2d\" (UID: \"efd5ed09-c17d-4aed-970a-4b85e5b1ef2d\") " Sep 30 14:20:43 crc kubenswrapper[4840]: I0930 14:20:43.051328 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grr2m\" (UniqueName: \"kubernetes.io/projected/efd5ed09-c17d-4aed-970a-4b85e5b1ef2d-kube-api-access-grr2m\") pod \"efd5ed09-c17d-4aed-970a-4b85e5b1ef2d\" (UID: \"efd5ed09-c17d-4aed-970a-4b85e5b1ef2d\") " Sep 30 14:20:43 crc kubenswrapper[4840]: I0930 14:20:43.052702 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efd5ed09-c17d-4aed-970a-4b85e5b1ef2d-utilities" (OuterVolumeSpecName: "utilities") pod "efd5ed09-c17d-4aed-970a-4b85e5b1ef2d" (UID: "efd5ed09-c17d-4aed-970a-4b85e5b1ef2d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:20:43 crc kubenswrapper[4840]: I0930 14:20:43.058423 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efd5ed09-c17d-4aed-970a-4b85e5b1ef2d-kube-api-access-grr2m" (OuterVolumeSpecName: "kube-api-access-grr2m") pod "efd5ed09-c17d-4aed-970a-4b85e5b1ef2d" (UID: "efd5ed09-c17d-4aed-970a-4b85e5b1ef2d"). InnerVolumeSpecName "kube-api-access-grr2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:20:43 crc kubenswrapper[4840]: I0930 14:20:43.096354 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efd5ed09-c17d-4aed-970a-4b85e5b1ef2d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "efd5ed09-c17d-4aed-970a-4b85e5b1ef2d" (UID: "efd5ed09-c17d-4aed-970a-4b85e5b1ef2d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:20:43 crc kubenswrapper[4840]: I0930 14:20:43.153906 4840 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efd5ed09-c17d-4aed-970a-4b85e5b1ef2d-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:20:43 crc kubenswrapper[4840]: I0930 14:20:43.153943 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grr2m\" (UniqueName: \"kubernetes.io/projected/efd5ed09-c17d-4aed-970a-4b85e5b1ef2d-kube-api-access-grr2m\") on node \"crc\" DevicePath \"\"" Sep 30 14:20:43 crc kubenswrapper[4840]: I0930 14:20:43.153953 4840 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efd5ed09-c17d-4aed-970a-4b85e5b1ef2d-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:20:43 crc kubenswrapper[4840]: I0930 14:20:43.608923 4840 generic.go:334] "Generic (PLEG): container finished" podID="5e5fc6b6-b432-4d70-8704-bbd2bd637691" containerID="f4ffdc4b3980f4c74ff4f9b818950753f5e6ae92cd9a4e3b8823bcd146774b4a" exitCode=0 Sep 30 14:20:43 crc kubenswrapper[4840]: I0930 14:20:43.608973 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m7mq8" event={"ID":"5e5fc6b6-b432-4d70-8704-bbd2bd637691","Type":"ContainerDied","Data":"f4ffdc4b3980f4c74ff4f9b818950753f5e6ae92cd9a4e3b8823bcd146774b4a"} Sep 30 14:20:43 crc kubenswrapper[4840]: I0930 14:20:43.611937 4840 generic.go:334] "Generic (PLEG): container finished" podID="efd5ed09-c17d-4aed-970a-4b85e5b1ef2d" containerID="2308dbee2c22a4ec51f06fdbe26bc009704eb00dc91446396f26dc6627bb3ed8" exitCode=0 Sep 30 14:20:43 crc kubenswrapper[4840]: I0930 14:20:43.611977 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rd7k5" event={"ID":"efd5ed09-c17d-4aed-970a-4b85e5b1ef2d","Type":"ContainerDied","Data":"2308dbee2c22a4ec51f06fdbe26bc009704eb00dc91446396f26dc6627bb3ed8"} Sep 30 14:20:43 crc kubenswrapper[4840]: I0930 14:20:43.612007 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rd7k5" event={"ID":"efd5ed09-c17d-4aed-970a-4b85e5b1ef2d","Type":"ContainerDied","Data":"c20eb480878f044f2e06b1b6f4be85bdfdee553c90cc57dca4c7d52b9cef31d0"} Sep 30 14:20:43 crc kubenswrapper[4840]: I0930 14:20:43.612022 4840 scope.go:117] "RemoveContainer" containerID="2308dbee2c22a4ec51f06fdbe26bc009704eb00dc91446396f26dc6627bb3ed8" Sep 30 14:20:43 crc kubenswrapper[4840]: I0930 14:20:43.612057 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rd7k5" Sep 30 14:20:43 crc kubenswrapper[4840]: I0930 14:20:43.642045 4840 scope.go:117] "RemoveContainer" containerID="8e4c26e67ac1d5d3719742bce870aef3b90f9db3f01d9ae0b0ea32ba884f743a" Sep 30 14:20:43 crc kubenswrapper[4840]: I0930 14:20:43.661565 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rd7k5"] Sep 30 14:20:43 crc kubenswrapper[4840]: I0930 14:20:43.671962 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rd7k5"] Sep 30 14:20:43 crc kubenswrapper[4840]: I0930 14:20:43.684750 4840 scope.go:117] "RemoveContainer" containerID="b947511904f690db585eb7e5bc7cf3ca431f429126a6fa7ee42c63d6a6fcbadb" Sep 30 14:20:43 crc kubenswrapper[4840]: I0930 14:20:43.722250 4840 scope.go:117] "RemoveContainer" containerID="2308dbee2c22a4ec51f06fdbe26bc009704eb00dc91446396f26dc6627bb3ed8" Sep 30 14:20:43 crc kubenswrapper[4840]: E0930 14:20:43.722869 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2308dbee2c22a4ec51f06fdbe26bc009704eb00dc91446396f26dc6627bb3ed8\": container with ID starting with 2308dbee2c22a4ec51f06fdbe26bc009704eb00dc91446396f26dc6627bb3ed8 not found: ID does not exist" containerID="2308dbee2c22a4ec51f06fdbe26bc009704eb00dc91446396f26dc6627bb3ed8" Sep 30 14:20:43 crc kubenswrapper[4840]: I0930 14:20:43.722913 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2308dbee2c22a4ec51f06fdbe26bc009704eb00dc91446396f26dc6627bb3ed8"} err="failed to get container status \"2308dbee2c22a4ec51f06fdbe26bc009704eb00dc91446396f26dc6627bb3ed8\": rpc error: code = NotFound desc = could not find container \"2308dbee2c22a4ec51f06fdbe26bc009704eb00dc91446396f26dc6627bb3ed8\": container with ID starting with 2308dbee2c22a4ec51f06fdbe26bc009704eb00dc91446396f26dc6627bb3ed8 not found: ID does not exist" Sep 30 14:20:43 crc kubenswrapper[4840]: I0930 14:20:43.722943 4840 scope.go:117] "RemoveContainer" containerID="8e4c26e67ac1d5d3719742bce870aef3b90f9db3f01d9ae0b0ea32ba884f743a" Sep 30 14:20:43 crc kubenswrapper[4840]: E0930 14:20:43.723467 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e4c26e67ac1d5d3719742bce870aef3b90f9db3f01d9ae0b0ea32ba884f743a\": container with ID starting with 8e4c26e67ac1d5d3719742bce870aef3b90f9db3f01d9ae0b0ea32ba884f743a not found: ID does not exist" containerID="8e4c26e67ac1d5d3719742bce870aef3b90f9db3f01d9ae0b0ea32ba884f743a" Sep 30 14:20:43 crc kubenswrapper[4840]: I0930 14:20:43.723527 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e4c26e67ac1d5d3719742bce870aef3b90f9db3f01d9ae0b0ea32ba884f743a"} err="failed to get container status \"8e4c26e67ac1d5d3719742bce870aef3b90f9db3f01d9ae0b0ea32ba884f743a\": rpc error: code = NotFound desc = could not find container \"8e4c26e67ac1d5d3719742bce870aef3b90f9db3f01d9ae0b0ea32ba884f743a\": container with ID starting with 8e4c26e67ac1d5d3719742bce870aef3b90f9db3f01d9ae0b0ea32ba884f743a not found: ID does not exist" Sep 30 14:20:43 crc kubenswrapper[4840]: I0930 14:20:43.723592 4840 scope.go:117] "RemoveContainer" containerID="b947511904f690db585eb7e5bc7cf3ca431f429126a6fa7ee42c63d6a6fcbadb" Sep 30 14:20:43 crc kubenswrapper[4840]: E0930 14:20:43.724065 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b947511904f690db585eb7e5bc7cf3ca431f429126a6fa7ee42c63d6a6fcbadb\": container with ID starting with b947511904f690db585eb7e5bc7cf3ca431f429126a6fa7ee42c63d6a6fcbadb not found: ID does not exist" containerID="b947511904f690db585eb7e5bc7cf3ca431f429126a6fa7ee42c63d6a6fcbadb" Sep 30 14:20:43 crc kubenswrapper[4840]: I0930 14:20:43.724116 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b947511904f690db585eb7e5bc7cf3ca431f429126a6fa7ee42c63d6a6fcbadb"} err="failed to get container status \"b947511904f690db585eb7e5bc7cf3ca431f429126a6fa7ee42c63d6a6fcbadb\": rpc error: code = NotFound desc = could not find container \"b947511904f690db585eb7e5bc7cf3ca431f429126a6fa7ee42c63d6a6fcbadb\": container with ID starting with b947511904f690db585eb7e5bc7cf3ca431f429126a6fa7ee42c63d6a6fcbadb not found: ID does not exist" Sep 30 14:20:44 crc kubenswrapper[4840]: I0930 14:20:44.127643 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efd5ed09-c17d-4aed-970a-4b85e5b1ef2d" path="/var/lib/kubelet/pods/efd5ed09-c17d-4aed-970a-4b85e5b1ef2d/volumes" Sep 30 14:20:44 crc kubenswrapper[4840]: I0930 14:20:44.627796 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m7mq8" event={"ID":"5e5fc6b6-b432-4d70-8704-bbd2bd637691","Type":"ContainerStarted","Data":"07fe3929eb29aa22bcf42b86b5285d53304a4d60e21718562a536bdf52f8bce1"} Sep 30 14:20:44 crc kubenswrapper[4840]: I0930 14:20:44.646976 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-m7mq8" podStartSLOduration=3.083614194 podStartE2EDuration="9.646956178s" podCreationTimestamp="2025-09-30 14:20:35 +0000 UTC" firstStartedPulling="2025-09-30 14:20:37.546094072 +0000 UTC m=+1466.175180505" lastFinishedPulling="2025-09-30 14:20:44.109436066 +0000 UTC m=+1472.738522489" observedRunningTime="2025-09-30 14:20:44.643409166 +0000 UTC m=+1473.272495589" watchObservedRunningTime="2025-09-30 14:20:44.646956178 +0000 UTC m=+1473.276042601" Sep 30 14:20:45 crc kubenswrapper[4840]: I0930 14:20:45.572679 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-m7mq8" Sep 30 14:20:45 crc kubenswrapper[4840]: I0930 14:20:45.573056 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-m7mq8" Sep 30 14:20:46 crc kubenswrapper[4840]: I0930 14:20:46.617418 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-m7mq8" podUID="5e5fc6b6-b432-4d70-8704-bbd2bd637691" containerName="registry-server" probeResult="failure" output=< Sep 30 14:20:46 crc kubenswrapper[4840]: timeout: failed to connect service ":50051" within 1s Sep 30 14:20:46 crc kubenswrapper[4840]: > Sep 30 14:20:51 crc kubenswrapper[4840]: I0930 14:20:51.871838 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:20:51 crc kubenswrapper[4840]: I0930 14:20:51.872436 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:20:51 crc kubenswrapper[4840]: I0930 14:20:51.872497 4840 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-747gk" Sep 30 14:20:51 crc kubenswrapper[4840]: I0930 14:20:51.873306 4840 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b4ec147091f38f805d294a67b19870b478b6bf3d54869eabd43f91f49541a76e"} pod="openshift-machine-config-operator/machine-config-daemon-747gk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 14:20:51 crc kubenswrapper[4840]: I0930 14:20:51.873378 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" containerID="cri-o://b4ec147091f38f805d294a67b19870b478b6bf3d54869eabd43f91f49541a76e" gracePeriod=600 Sep 30 14:20:52 crc kubenswrapper[4840]: I0930 14:20:52.699460 4840 generic.go:334] "Generic (PLEG): container finished" podID="10e8b890-7f20-4a36-8e03-898620cf599a" containerID="b4ec147091f38f805d294a67b19870b478b6bf3d54869eabd43f91f49541a76e" exitCode=0 Sep 30 14:20:52 crc kubenswrapper[4840]: I0930 14:20:52.699518 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" event={"ID":"10e8b890-7f20-4a36-8e03-898620cf599a","Type":"ContainerDied","Data":"b4ec147091f38f805d294a67b19870b478b6bf3d54869eabd43f91f49541a76e"} Sep 30 14:20:52 crc kubenswrapper[4840]: I0930 14:20:52.700006 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" event={"ID":"10e8b890-7f20-4a36-8e03-898620cf599a","Type":"ContainerStarted","Data":"74b12cc9318ac99161f1b2541861d89b6100ba9a986e5863f53ae26c2a8953e7"} Sep 30 14:20:52 crc kubenswrapper[4840]: I0930 14:20:52.700023 4840 scope.go:117] "RemoveContainer" containerID="5170b28c065e5cc00d1164411d96a8eb01e1860be6b5701def2ab76e6fb63860" Sep 30 14:20:53 crc kubenswrapper[4840]: I0930 14:20:53.712196 4840 generic.go:334] "Generic (PLEG): container finished" podID="558c75bb-7947-423d-b3aa-1b2ab4ba2d1d" containerID="f5a2f1cea1343ab9cbdd93d48463fd9a34407ef0712526e2b7a909c4919b614e" exitCode=0 Sep 30 14:20:53 crc kubenswrapper[4840]: I0930 14:20:53.712301 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6ntwv" event={"ID":"558c75bb-7947-423d-b3aa-1b2ab4ba2d1d","Type":"ContainerDied","Data":"f5a2f1cea1343ab9cbdd93d48463fd9a34407ef0712526e2b7a909c4919b614e"} Sep 30 14:20:55 crc kubenswrapper[4840]: I0930 14:20:55.133007 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6ntwv" Sep 30 14:20:55 crc kubenswrapper[4840]: I0930 14:20:55.184454 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/558c75bb-7947-423d-b3aa-1b2ab4ba2d1d-bootstrap-combined-ca-bundle\") pod \"558c75bb-7947-423d-b3aa-1b2ab4ba2d1d\" (UID: \"558c75bb-7947-423d-b3aa-1b2ab4ba2d1d\") " Sep 30 14:20:55 crc kubenswrapper[4840]: I0930 14:20:55.184576 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/558c75bb-7947-423d-b3aa-1b2ab4ba2d1d-inventory\") pod \"558c75bb-7947-423d-b3aa-1b2ab4ba2d1d\" (UID: \"558c75bb-7947-423d-b3aa-1b2ab4ba2d1d\") " Sep 30 14:20:55 crc kubenswrapper[4840]: I0930 14:20:55.184600 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbpqb\" (UniqueName: \"kubernetes.io/projected/558c75bb-7947-423d-b3aa-1b2ab4ba2d1d-kube-api-access-mbpqb\") pod \"558c75bb-7947-423d-b3aa-1b2ab4ba2d1d\" (UID: \"558c75bb-7947-423d-b3aa-1b2ab4ba2d1d\") " Sep 30 14:20:55 crc kubenswrapper[4840]: I0930 14:20:55.185525 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/558c75bb-7947-423d-b3aa-1b2ab4ba2d1d-ssh-key\") pod \"558c75bb-7947-423d-b3aa-1b2ab4ba2d1d\" (UID: \"558c75bb-7947-423d-b3aa-1b2ab4ba2d1d\") " Sep 30 14:20:55 crc kubenswrapper[4840]: I0930 14:20:55.191691 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/558c75bb-7947-423d-b3aa-1b2ab4ba2d1d-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "558c75bb-7947-423d-b3aa-1b2ab4ba2d1d" (UID: "558c75bb-7947-423d-b3aa-1b2ab4ba2d1d"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:20:55 crc kubenswrapper[4840]: I0930 14:20:55.191689 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/558c75bb-7947-423d-b3aa-1b2ab4ba2d1d-kube-api-access-mbpqb" (OuterVolumeSpecName: "kube-api-access-mbpqb") pod "558c75bb-7947-423d-b3aa-1b2ab4ba2d1d" (UID: "558c75bb-7947-423d-b3aa-1b2ab4ba2d1d"). InnerVolumeSpecName "kube-api-access-mbpqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:20:55 crc kubenswrapper[4840]: I0930 14:20:55.214298 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/558c75bb-7947-423d-b3aa-1b2ab4ba2d1d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "558c75bb-7947-423d-b3aa-1b2ab4ba2d1d" (UID: "558c75bb-7947-423d-b3aa-1b2ab4ba2d1d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:20:55 crc kubenswrapper[4840]: I0930 14:20:55.214782 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/558c75bb-7947-423d-b3aa-1b2ab4ba2d1d-inventory" (OuterVolumeSpecName: "inventory") pod "558c75bb-7947-423d-b3aa-1b2ab4ba2d1d" (UID: "558c75bb-7947-423d-b3aa-1b2ab4ba2d1d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:20:55 crc kubenswrapper[4840]: I0930 14:20:55.289305 4840 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/558c75bb-7947-423d-b3aa-1b2ab4ba2d1d-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:20:55 crc kubenswrapper[4840]: I0930 14:20:55.289389 4840 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/558c75bb-7947-423d-b3aa-1b2ab4ba2d1d-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 14:20:55 crc kubenswrapper[4840]: I0930 14:20:55.289406 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbpqb\" (UniqueName: \"kubernetes.io/projected/558c75bb-7947-423d-b3aa-1b2ab4ba2d1d-kube-api-access-mbpqb\") on node \"crc\" DevicePath \"\"" Sep 30 14:20:55 crc kubenswrapper[4840]: I0930 14:20:55.289418 4840 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/558c75bb-7947-423d-b3aa-1b2ab4ba2d1d-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:20:55 crc kubenswrapper[4840]: I0930 14:20:55.624816 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-m7mq8" Sep 30 14:20:55 crc kubenswrapper[4840]: I0930 14:20:55.678463 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-m7mq8" Sep 30 14:20:55 crc kubenswrapper[4840]: I0930 14:20:55.734601 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6ntwv" event={"ID":"558c75bb-7947-423d-b3aa-1b2ab4ba2d1d","Type":"ContainerDied","Data":"0df5995a0b3eb6a76478cd96d84419785237ae2ff95b871f1ffdc1e510433c10"} Sep 30 14:20:55 crc kubenswrapper[4840]: I0930 14:20:55.734649 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0df5995a0b3eb6a76478cd96d84419785237ae2ff95b871f1ffdc1e510433c10" Sep 30 14:20:55 crc kubenswrapper[4840]: I0930 14:20:55.734661 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6ntwv" Sep 30 14:20:55 crc kubenswrapper[4840]: I0930 14:20:55.810603 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cbh6n"] Sep 30 14:20:55 crc kubenswrapper[4840]: E0930 14:20:55.811078 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efd5ed09-c17d-4aed-970a-4b85e5b1ef2d" containerName="extract-content" Sep 30 14:20:55 crc kubenswrapper[4840]: I0930 14:20:55.811100 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="efd5ed09-c17d-4aed-970a-4b85e5b1ef2d" containerName="extract-content" Sep 30 14:20:55 crc kubenswrapper[4840]: E0930 14:20:55.811113 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efd5ed09-c17d-4aed-970a-4b85e5b1ef2d" containerName="registry-server" Sep 30 14:20:55 crc kubenswrapper[4840]: I0930 14:20:55.811124 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="efd5ed09-c17d-4aed-970a-4b85e5b1ef2d" containerName="registry-server" Sep 30 14:20:55 crc kubenswrapper[4840]: E0930 14:20:55.811139 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="558c75bb-7947-423d-b3aa-1b2ab4ba2d1d" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Sep 30 14:20:55 crc kubenswrapper[4840]: I0930 14:20:55.811150 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="558c75bb-7947-423d-b3aa-1b2ab4ba2d1d" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Sep 30 14:20:55 crc kubenswrapper[4840]: E0930 14:20:55.811188 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efd5ed09-c17d-4aed-970a-4b85e5b1ef2d" containerName="extract-utilities" Sep 30 14:20:55 crc kubenswrapper[4840]: I0930 14:20:55.811198 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="efd5ed09-c17d-4aed-970a-4b85e5b1ef2d" containerName="extract-utilities" Sep 30 14:20:55 crc kubenswrapper[4840]: I0930 14:20:55.811423 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="558c75bb-7947-423d-b3aa-1b2ab4ba2d1d" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Sep 30 14:20:55 crc kubenswrapper[4840]: I0930 14:20:55.811457 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="efd5ed09-c17d-4aed-970a-4b85e5b1ef2d" containerName="registry-server" Sep 30 14:20:55 crc kubenswrapper[4840]: I0930 14:20:55.812241 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cbh6n" Sep 30 14:20:55 crc kubenswrapper[4840]: I0930 14:20:55.814754 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 14:20:55 crc kubenswrapper[4840]: I0930 14:20:55.815120 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9zm9z" Sep 30 14:20:55 crc kubenswrapper[4840]: I0930 14:20:55.815956 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 14:20:55 crc kubenswrapper[4840]: I0930 14:20:55.817604 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 14:20:55 crc kubenswrapper[4840]: I0930 14:20:55.823473 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cbh6n"] Sep 30 14:20:55 crc kubenswrapper[4840]: I0930 14:20:55.860312 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-m7mq8"] Sep 30 14:20:55 crc kubenswrapper[4840]: I0930 14:20:55.900360 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wq8w9\" (UniqueName: \"kubernetes.io/projected/17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780-kube-api-access-wq8w9\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cbh6n\" (UID: \"17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cbh6n" Sep 30 14:20:55 crc kubenswrapper[4840]: I0930 14:20:55.900423 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cbh6n\" (UID: \"17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cbh6n" Sep 30 14:20:55 crc kubenswrapper[4840]: I0930 14:20:55.900802 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cbh6n\" (UID: \"17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cbh6n" Sep 30 14:20:56 crc kubenswrapper[4840]: I0930 14:20:56.002935 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cbh6n\" (UID: \"17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cbh6n" Sep 30 14:20:56 crc kubenswrapper[4840]: I0930 14:20:56.003037 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wq8w9\" (UniqueName: \"kubernetes.io/projected/17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780-kube-api-access-wq8w9\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cbh6n\" (UID: \"17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cbh6n" Sep 30 14:20:56 crc kubenswrapper[4840]: I0930 14:20:56.003071 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cbh6n\" (UID: \"17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cbh6n" Sep 30 14:20:56 crc kubenswrapper[4840]: I0930 14:20:56.007312 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cbh6n\" (UID: \"17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cbh6n" Sep 30 14:20:56 crc kubenswrapper[4840]: I0930 14:20:56.008765 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cbh6n\" (UID: \"17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cbh6n" Sep 30 14:20:56 crc kubenswrapper[4840]: I0930 14:20:56.021209 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wq8w9\" (UniqueName: \"kubernetes.io/projected/17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780-kube-api-access-wq8w9\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cbh6n\" (UID: \"17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cbh6n" Sep 30 14:20:56 crc kubenswrapper[4840]: I0930 14:20:56.130044 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cbh6n" Sep 30 14:20:56 crc kubenswrapper[4840]: I0930 14:20:56.628310 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cbh6n"] Sep 30 14:20:56 crc kubenswrapper[4840]: I0930 14:20:56.746900 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cbh6n" event={"ID":"17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780","Type":"ContainerStarted","Data":"d613834e387427b7f72266a2a11981620cd50dc20fb82f2fc93c37ab7a37b681"} Sep 30 14:20:56 crc kubenswrapper[4840]: I0930 14:20:56.747082 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-m7mq8" podUID="5e5fc6b6-b432-4d70-8704-bbd2bd637691" containerName="registry-server" containerID="cri-o://07fe3929eb29aa22bcf42b86b5285d53304a4d60e21718562a536bdf52f8bce1" gracePeriod=2 Sep 30 14:20:57 crc kubenswrapper[4840]: I0930 14:20:57.233912 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m7mq8" Sep 30 14:20:57 crc kubenswrapper[4840]: I0930 14:20:57.325251 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r899z\" (UniqueName: \"kubernetes.io/projected/5e5fc6b6-b432-4d70-8704-bbd2bd637691-kube-api-access-r899z\") pod \"5e5fc6b6-b432-4d70-8704-bbd2bd637691\" (UID: \"5e5fc6b6-b432-4d70-8704-bbd2bd637691\") " Sep 30 14:20:57 crc kubenswrapper[4840]: I0930 14:20:57.325334 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e5fc6b6-b432-4d70-8704-bbd2bd637691-catalog-content\") pod \"5e5fc6b6-b432-4d70-8704-bbd2bd637691\" (UID: \"5e5fc6b6-b432-4d70-8704-bbd2bd637691\") " Sep 30 14:20:57 crc kubenswrapper[4840]: I0930 14:20:57.325520 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e5fc6b6-b432-4d70-8704-bbd2bd637691-utilities\") pod \"5e5fc6b6-b432-4d70-8704-bbd2bd637691\" (UID: \"5e5fc6b6-b432-4d70-8704-bbd2bd637691\") " Sep 30 14:20:57 crc kubenswrapper[4840]: I0930 14:20:57.326813 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e5fc6b6-b432-4d70-8704-bbd2bd637691-utilities" (OuterVolumeSpecName: "utilities") pod "5e5fc6b6-b432-4d70-8704-bbd2bd637691" (UID: "5e5fc6b6-b432-4d70-8704-bbd2bd637691"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:20:57 crc kubenswrapper[4840]: I0930 14:20:57.331186 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e5fc6b6-b432-4d70-8704-bbd2bd637691-kube-api-access-r899z" (OuterVolumeSpecName: "kube-api-access-r899z") pod "5e5fc6b6-b432-4d70-8704-bbd2bd637691" (UID: "5e5fc6b6-b432-4d70-8704-bbd2bd637691"). InnerVolumeSpecName "kube-api-access-r899z". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:20:57 crc kubenswrapper[4840]: I0930 14:20:57.425989 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e5fc6b6-b432-4d70-8704-bbd2bd637691-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5e5fc6b6-b432-4d70-8704-bbd2bd637691" (UID: "5e5fc6b6-b432-4d70-8704-bbd2bd637691"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:20:57 crc kubenswrapper[4840]: I0930 14:20:57.426875 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e5fc6b6-b432-4d70-8704-bbd2bd637691-catalog-content\") pod \"5e5fc6b6-b432-4d70-8704-bbd2bd637691\" (UID: \"5e5fc6b6-b432-4d70-8704-bbd2bd637691\") " Sep 30 14:20:57 crc kubenswrapper[4840]: W0930 14:20:57.426995 4840 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/5e5fc6b6-b432-4d70-8704-bbd2bd637691/volumes/kubernetes.io~empty-dir/catalog-content Sep 30 14:20:57 crc kubenswrapper[4840]: I0930 14:20:57.427028 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e5fc6b6-b432-4d70-8704-bbd2bd637691-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5e5fc6b6-b432-4d70-8704-bbd2bd637691" (UID: "5e5fc6b6-b432-4d70-8704-bbd2bd637691"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:20:57 crc kubenswrapper[4840]: I0930 14:20:57.427565 4840 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e5fc6b6-b432-4d70-8704-bbd2bd637691-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:20:57 crc kubenswrapper[4840]: I0930 14:20:57.427599 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r899z\" (UniqueName: \"kubernetes.io/projected/5e5fc6b6-b432-4d70-8704-bbd2bd637691-kube-api-access-r899z\") on node \"crc\" DevicePath \"\"" Sep 30 14:20:57 crc kubenswrapper[4840]: I0930 14:20:57.427617 4840 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e5fc6b6-b432-4d70-8704-bbd2bd637691-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:20:57 crc kubenswrapper[4840]: I0930 14:20:57.757752 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cbh6n" event={"ID":"17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780","Type":"ContainerStarted","Data":"3ac453278ca8bea6b41fdb8b32e7868671ba999c75e15dc701216a1b307bf758"} Sep 30 14:20:57 crc kubenswrapper[4840]: I0930 14:20:57.760011 4840 generic.go:334] "Generic (PLEG): container finished" podID="5e5fc6b6-b432-4d70-8704-bbd2bd637691" containerID="07fe3929eb29aa22bcf42b86b5285d53304a4d60e21718562a536bdf52f8bce1" exitCode=0 Sep 30 14:20:57 crc kubenswrapper[4840]: I0930 14:20:57.760031 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m7mq8" Sep 30 14:20:57 crc kubenswrapper[4840]: I0930 14:20:57.760048 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m7mq8" event={"ID":"5e5fc6b6-b432-4d70-8704-bbd2bd637691","Type":"ContainerDied","Data":"07fe3929eb29aa22bcf42b86b5285d53304a4d60e21718562a536bdf52f8bce1"} Sep 30 14:20:57 crc kubenswrapper[4840]: I0930 14:20:57.760126 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m7mq8" event={"ID":"5e5fc6b6-b432-4d70-8704-bbd2bd637691","Type":"ContainerDied","Data":"88bed3eebd86db9b64af492f6ea3ce784bc70f31127bb30ba6dccc41bc78f68a"} Sep 30 14:20:57 crc kubenswrapper[4840]: I0930 14:20:57.760156 4840 scope.go:117] "RemoveContainer" containerID="07fe3929eb29aa22bcf42b86b5285d53304a4d60e21718562a536bdf52f8bce1" Sep 30 14:20:57 crc kubenswrapper[4840]: I0930 14:20:57.783976 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cbh6n" podStartSLOduration=2.296067871 podStartE2EDuration="2.783959773s" podCreationTimestamp="2025-09-30 14:20:55 +0000 UTC" firstStartedPulling="2025-09-30 14:20:56.634274445 +0000 UTC m=+1485.263360868" lastFinishedPulling="2025-09-30 14:20:57.122166347 +0000 UTC m=+1485.751252770" observedRunningTime="2025-09-30 14:20:57.77377099 +0000 UTC m=+1486.402857413" watchObservedRunningTime="2025-09-30 14:20:57.783959773 +0000 UTC m=+1486.413046196" Sep 30 14:20:57 crc kubenswrapper[4840]: I0930 14:20:57.802505 4840 scope.go:117] "RemoveContainer" containerID="f4ffdc4b3980f4c74ff4f9b818950753f5e6ae92cd9a4e3b8823bcd146774b4a" Sep 30 14:20:57 crc kubenswrapper[4840]: I0930 14:20:57.813542 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-m7mq8"] Sep 30 14:20:57 crc kubenswrapper[4840]: I0930 14:20:57.822123 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-m7mq8"] Sep 30 14:20:57 crc kubenswrapper[4840]: I0930 14:20:57.840744 4840 scope.go:117] "RemoveContainer" containerID="c8c3fd0e2a9d08ae30b12f9007523d299027aada67ddae87e9a128ad08943329" Sep 30 14:20:57 crc kubenswrapper[4840]: I0930 14:20:57.880774 4840 scope.go:117] "RemoveContainer" containerID="07fe3929eb29aa22bcf42b86b5285d53304a4d60e21718562a536bdf52f8bce1" Sep 30 14:20:57 crc kubenswrapper[4840]: E0930 14:20:57.881356 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07fe3929eb29aa22bcf42b86b5285d53304a4d60e21718562a536bdf52f8bce1\": container with ID starting with 07fe3929eb29aa22bcf42b86b5285d53304a4d60e21718562a536bdf52f8bce1 not found: ID does not exist" containerID="07fe3929eb29aa22bcf42b86b5285d53304a4d60e21718562a536bdf52f8bce1" Sep 30 14:20:57 crc kubenswrapper[4840]: I0930 14:20:57.881413 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07fe3929eb29aa22bcf42b86b5285d53304a4d60e21718562a536bdf52f8bce1"} err="failed to get container status \"07fe3929eb29aa22bcf42b86b5285d53304a4d60e21718562a536bdf52f8bce1\": rpc error: code = NotFound desc = could not find container \"07fe3929eb29aa22bcf42b86b5285d53304a4d60e21718562a536bdf52f8bce1\": container with ID starting with 07fe3929eb29aa22bcf42b86b5285d53304a4d60e21718562a536bdf52f8bce1 not found: ID does not exist" Sep 30 14:20:57 crc kubenswrapper[4840]: I0930 14:20:57.881447 4840 scope.go:117] "RemoveContainer" containerID="f4ffdc4b3980f4c74ff4f9b818950753f5e6ae92cd9a4e3b8823bcd146774b4a" Sep 30 14:20:57 crc kubenswrapper[4840]: E0930 14:20:57.881899 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4ffdc4b3980f4c74ff4f9b818950753f5e6ae92cd9a4e3b8823bcd146774b4a\": container with ID starting with f4ffdc4b3980f4c74ff4f9b818950753f5e6ae92cd9a4e3b8823bcd146774b4a not found: ID does not exist" containerID="f4ffdc4b3980f4c74ff4f9b818950753f5e6ae92cd9a4e3b8823bcd146774b4a" Sep 30 14:20:57 crc kubenswrapper[4840]: I0930 14:20:57.881928 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4ffdc4b3980f4c74ff4f9b818950753f5e6ae92cd9a4e3b8823bcd146774b4a"} err="failed to get container status \"f4ffdc4b3980f4c74ff4f9b818950753f5e6ae92cd9a4e3b8823bcd146774b4a\": rpc error: code = NotFound desc = could not find container \"f4ffdc4b3980f4c74ff4f9b818950753f5e6ae92cd9a4e3b8823bcd146774b4a\": container with ID starting with f4ffdc4b3980f4c74ff4f9b818950753f5e6ae92cd9a4e3b8823bcd146774b4a not found: ID does not exist" Sep 30 14:20:57 crc kubenswrapper[4840]: I0930 14:20:57.881946 4840 scope.go:117] "RemoveContainer" containerID="c8c3fd0e2a9d08ae30b12f9007523d299027aada67ddae87e9a128ad08943329" Sep 30 14:20:57 crc kubenswrapper[4840]: E0930 14:20:57.882332 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8c3fd0e2a9d08ae30b12f9007523d299027aada67ddae87e9a128ad08943329\": container with ID starting with c8c3fd0e2a9d08ae30b12f9007523d299027aada67ddae87e9a128ad08943329 not found: ID does not exist" containerID="c8c3fd0e2a9d08ae30b12f9007523d299027aada67ddae87e9a128ad08943329" Sep 30 14:20:57 crc kubenswrapper[4840]: I0930 14:20:57.882358 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8c3fd0e2a9d08ae30b12f9007523d299027aada67ddae87e9a128ad08943329"} err="failed to get container status \"c8c3fd0e2a9d08ae30b12f9007523d299027aada67ddae87e9a128ad08943329\": rpc error: code = NotFound desc = could not find container \"c8c3fd0e2a9d08ae30b12f9007523d299027aada67ddae87e9a128ad08943329\": container with ID starting with c8c3fd0e2a9d08ae30b12f9007523d299027aada67ddae87e9a128ad08943329 not found: ID does not exist" Sep 30 14:20:58 crc kubenswrapper[4840]: I0930 14:20:58.126585 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e5fc6b6-b432-4d70-8704-bbd2bd637691" path="/var/lib/kubelet/pods/5e5fc6b6-b432-4d70-8704-bbd2bd637691/volumes" Sep 30 14:21:33 crc kubenswrapper[4840]: I0930 14:21:33.265683 4840 scope.go:117] "RemoveContainer" containerID="4fa5e7cb35fa61ade81028ac2dd0b73d7848ab07034be304329ea1bfb9d0bc25" Sep 30 14:21:33 crc kubenswrapper[4840]: I0930 14:21:33.301153 4840 scope.go:117] "RemoveContainer" containerID="658f57946407b68ab2e09b557a0c2341838801cfbfb57c5cd469301197d45699" Sep 30 14:21:33 crc kubenswrapper[4840]: I0930 14:21:33.361853 4840 scope.go:117] "RemoveContainer" containerID="4cc932978f9822386e3a5f8f4135308469aed6b1b0a4f69f8c362a02759d7ab6" Sep 30 14:21:33 crc kubenswrapper[4840]: I0930 14:21:33.384599 4840 scope.go:117] "RemoveContainer" containerID="abf8546fd8c0c1e84bbe02106ed1287e7966858e265c9a54580b44162ded1c2f" Sep 30 14:21:33 crc kubenswrapper[4840]: I0930 14:21:33.410908 4840 scope.go:117] "RemoveContainer" containerID="9944d1240049ca763bd1030a91e7bfc104f71e03553fc6b95c95cf3a385e5784" Sep 30 14:21:33 crc kubenswrapper[4840]: I0930 14:21:33.432021 4840 scope.go:117] "RemoveContainer" containerID="274d7e35969926c142e97bdd427d59044bf48e96015d9626c7a3b9a3b23ee7a9" Sep 30 14:21:53 crc kubenswrapper[4840]: I0930 14:21:53.036156 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-n7hj6"] Sep 30 14:21:53 crc kubenswrapper[4840]: I0930 14:21:53.044327 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-n7hj6"] Sep 30 14:21:54 crc kubenswrapper[4840]: I0930 14:21:54.132725 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abd36ef0-abb9-4812-adcb-f3b56172b81e" path="/var/lib/kubelet/pods/abd36ef0-abb9-4812-adcb-f3b56172b81e/volumes" Sep 30 14:21:57 crc kubenswrapper[4840]: I0930 14:21:57.024591 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-g7vsc"] Sep 30 14:21:57 crc kubenswrapper[4840]: I0930 14:21:57.032382 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-g7vsc"] Sep 30 14:21:58 crc kubenswrapper[4840]: I0930 14:21:58.042077 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-5mfh2"] Sep 30 14:21:58 crc kubenswrapper[4840]: I0930 14:21:58.053313 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-5mfh2"] Sep 30 14:21:58 crc kubenswrapper[4840]: I0930 14:21:58.128956 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46aeb27d-7764-4fbb-bbdb-88628a7d6f59" path="/var/lib/kubelet/pods/46aeb27d-7764-4fbb-bbdb-88628a7d6f59/volumes" Sep 30 14:21:58 crc kubenswrapper[4840]: I0930 14:21:58.129842 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9bb72efc-899f-4e40-a519-86402ed702b7" path="/var/lib/kubelet/pods/9bb72efc-899f-4e40-a519-86402ed702b7/volumes" Sep 30 14:22:05 crc kubenswrapper[4840]: I0930 14:22:05.051621 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-ae90-account-create-4schj"] Sep 30 14:22:05 crc kubenswrapper[4840]: I0930 14:22:05.060497 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-ae90-account-create-4schj"] Sep 30 14:22:06 crc kubenswrapper[4840]: I0930 14:22:06.133929 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5f49283-d818-4180-9b52-4f05352586e4" path="/var/lib/kubelet/pods/b5f49283-d818-4180-9b52-4f05352586e4/volumes" Sep 30 14:22:07 crc kubenswrapper[4840]: I0930 14:22:07.037418 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-1a60-account-create-wnsbv"] Sep 30 14:22:07 crc kubenswrapper[4840]: I0930 14:22:07.046283 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-1a60-account-create-wnsbv"] Sep 30 14:22:08 crc kubenswrapper[4840]: I0930 14:22:08.026534 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-4e44-account-create-dcfn6"] Sep 30 14:22:08 crc kubenswrapper[4840]: I0930 14:22:08.033406 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-4e44-account-create-dcfn6"] Sep 30 14:22:08 crc kubenswrapper[4840]: I0930 14:22:08.129304 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36cd5d4f-6ef2-440f-a9b3-56610a475114" path="/var/lib/kubelet/pods/36cd5d4f-6ef2-440f-a9b3-56610a475114/volumes" Sep 30 14:22:08 crc kubenswrapper[4840]: I0930 14:22:08.130954 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc1ccb29-1c0d-40a2-ae2e-c277fff08748" path="/var/lib/kubelet/pods/fc1ccb29-1c0d-40a2-ae2e-c277fff08748/volumes" Sep 30 14:22:26 crc kubenswrapper[4840]: I0930 14:22:26.037875 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-g9msc"] Sep 30 14:22:26 crc kubenswrapper[4840]: I0930 14:22:26.048121 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-92cq6"] Sep 30 14:22:26 crc kubenswrapper[4840]: I0930 14:22:26.058950 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-kcv4b"] Sep 30 14:22:26 crc kubenswrapper[4840]: I0930 14:22:26.069525 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-92cq6"] Sep 30 14:22:26 crc kubenswrapper[4840]: I0930 14:22:26.077348 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-g9msc"] Sep 30 14:22:26 crc kubenswrapper[4840]: I0930 14:22:26.084718 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-kcv4b"] Sep 30 14:22:26 crc kubenswrapper[4840]: I0930 14:22:26.127419 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1012193a-00bc-4ffa-81ae-943299c3ac15" path="/var/lib/kubelet/pods/1012193a-00bc-4ffa-81ae-943299c3ac15/volumes" Sep 30 14:22:26 crc kubenswrapper[4840]: I0930 14:22:26.128068 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47747e51-6cb7-4944-a2a1-8011dc25a43a" path="/var/lib/kubelet/pods/47747e51-6cb7-4944-a2a1-8011dc25a43a/volumes" Sep 30 14:22:26 crc kubenswrapper[4840]: I0930 14:22:26.128586 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57b18c4f-a2dc-433c-96e0-fbcebd4ecb32" path="/var/lib/kubelet/pods/57b18c4f-a2dc-433c-96e0-fbcebd4ecb32/volumes" Sep 30 14:22:32 crc kubenswrapper[4840]: I0930 14:22:32.682184 4840 generic.go:334] "Generic (PLEG): container finished" podID="17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780" containerID="3ac453278ca8bea6b41fdb8b32e7868671ba999c75e15dc701216a1b307bf758" exitCode=0 Sep 30 14:22:32 crc kubenswrapper[4840]: I0930 14:22:32.682412 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cbh6n" event={"ID":"17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780","Type":"ContainerDied","Data":"3ac453278ca8bea6b41fdb8b32e7868671ba999c75e15dc701216a1b307bf758"} Sep 30 14:22:33 crc kubenswrapper[4840]: I0930 14:22:33.518985 4840 scope.go:117] "RemoveContainer" containerID="7cabc82331616339c77a20dabef3a6e5391ec674437f89c761356362e795e92c" Sep 30 14:22:33 crc kubenswrapper[4840]: I0930 14:22:33.546378 4840 scope.go:117] "RemoveContainer" containerID="493c705051312d1d93803dae118e3980ab83642f89ca3f4ed58c13e68c039320" Sep 30 14:22:33 crc kubenswrapper[4840]: I0930 14:22:33.604753 4840 scope.go:117] "RemoveContainer" containerID="a117172c136b13b2d7a0bafe4c4e9077993f24440d861fac8cb69865570a47e9" Sep 30 14:22:33 crc kubenswrapper[4840]: I0930 14:22:33.650610 4840 scope.go:117] "RemoveContainer" containerID="5cd9591b125dcec7aba5800e5f1bfd0c285ca6e095931a0566197f1f2764c825" Sep 30 14:22:33 crc kubenswrapper[4840]: I0930 14:22:33.695230 4840 scope.go:117] "RemoveContainer" containerID="55ce9a2b8c0bde8dcaaa1589f64cea57c71d079ce974061d3d2a6b9e5fae8594" Sep 30 14:22:33 crc kubenswrapper[4840]: I0930 14:22:33.747724 4840 scope.go:117] "RemoveContainer" containerID="26e779802fe5efa670adadddea2105ace90b73864327733e492daf4ef27f174d" Sep 30 14:22:33 crc kubenswrapper[4840]: I0930 14:22:33.783739 4840 scope.go:117] "RemoveContainer" containerID="6a13fa59195a1222469f537a473de6efef9f2bfdeb3434d22401e1c2abc9e2b6" Sep 30 14:22:33 crc kubenswrapper[4840]: I0930 14:22:33.803833 4840 scope.go:117] "RemoveContainer" containerID="2d128c0e805ecf2486bb4310324e54729f86c76e69942a710a8ba7dafedf57d5" Sep 30 14:22:33 crc kubenswrapper[4840]: I0930 14:22:33.830759 4840 scope.go:117] "RemoveContainer" containerID="ed388f6da8a23eac58e1d06b8c8262fb4e648b50383158e129180e1f203a3276" Sep 30 14:22:33 crc kubenswrapper[4840]: I0930 14:22:33.852428 4840 scope.go:117] "RemoveContainer" containerID="5bac448d393c5d2cdabbf98690a456182f58a8e6161c29658b38fb96e3b3bea1" Sep 30 14:22:33 crc kubenswrapper[4840]: I0930 14:22:33.871216 4840 scope.go:117] "RemoveContainer" containerID="1f59be2c7ab21244947c4b598d95af9dff7a71a384e668a7d6d375e330d18043" Sep 30 14:22:33 crc kubenswrapper[4840]: I0930 14:22:33.997407 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cbh6n" Sep 30 14:22:34 crc kubenswrapper[4840]: I0930 14:22:34.045935 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-czdht"] Sep 30 14:22:34 crc kubenswrapper[4840]: I0930 14:22:34.059095 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-czdht"] Sep 30 14:22:34 crc kubenswrapper[4840]: I0930 14:22:34.121543 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780-ssh-key\") pod \"17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780\" (UID: \"17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780\") " Sep 30 14:22:34 crc kubenswrapper[4840]: I0930 14:22:34.121827 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780-inventory\") pod \"17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780\" (UID: \"17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780\") " Sep 30 14:22:34 crc kubenswrapper[4840]: I0930 14:22:34.121888 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wq8w9\" (UniqueName: \"kubernetes.io/projected/17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780-kube-api-access-wq8w9\") pod \"17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780\" (UID: \"17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780\") " Sep 30 14:22:34 crc kubenswrapper[4840]: I0930 14:22:34.127878 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780-kube-api-access-wq8w9" (OuterVolumeSpecName: "kube-api-access-wq8w9") pod "17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780" (UID: "17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780"). InnerVolumeSpecName "kube-api-access-wq8w9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:22:34 crc kubenswrapper[4840]: I0930 14:22:34.129002 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1f15dc9-8926-4c0c-89da-a27216c4cb1a" path="/var/lib/kubelet/pods/a1f15dc9-8926-4c0c-89da-a27216c4cb1a/volumes" Sep 30 14:22:34 crc kubenswrapper[4840]: I0930 14:22:34.150347 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780-inventory" (OuterVolumeSpecName: "inventory") pod "17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780" (UID: "17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:22:34 crc kubenswrapper[4840]: I0930 14:22:34.150806 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780" (UID: "17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:22:34 crc kubenswrapper[4840]: I0930 14:22:34.224715 4840 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:22:34 crc kubenswrapper[4840]: I0930 14:22:34.224814 4840 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 14:22:34 crc kubenswrapper[4840]: I0930 14:22:34.224841 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wq8w9\" (UniqueName: \"kubernetes.io/projected/17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780-kube-api-access-wq8w9\") on node \"crc\" DevicePath \"\"" Sep 30 14:22:34 crc kubenswrapper[4840]: I0930 14:22:34.710654 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cbh6n" event={"ID":"17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780","Type":"ContainerDied","Data":"d613834e387427b7f72266a2a11981620cd50dc20fb82f2fc93c37ab7a37b681"} Sep 30 14:22:34 crc kubenswrapper[4840]: I0930 14:22:34.710695 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d613834e387427b7f72266a2a11981620cd50dc20fb82f2fc93c37ab7a37b681" Sep 30 14:22:34 crc kubenswrapper[4840]: I0930 14:22:34.710759 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cbh6n" Sep 30 14:22:34 crc kubenswrapper[4840]: I0930 14:22:34.782457 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xvhf8"] Sep 30 14:22:34 crc kubenswrapper[4840]: E0930 14:22:34.782879 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e5fc6b6-b432-4d70-8704-bbd2bd637691" containerName="extract-content" Sep 30 14:22:34 crc kubenswrapper[4840]: I0930 14:22:34.782902 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e5fc6b6-b432-4d70-8704-bbd2bd637691" containerName="extract-content" Sep 30 14:22:34 crc kubenswrapper[4840]: E0930 14:22:34.782956 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e5fc6b6-b432-4d70-8704-bbd2bd637691" containerName="registry-server" Sep 30 14:22:34 crc kubenswrapper[4840]: I0930 14:22:34.782965 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e5fc6b6-b432-4d70-8704-bbd2bd637691" containerName="registry-server" Sep 30 14:22:34 crc kubenswrapper[4840]: E0930 14:22:34.782983 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e5fc6b6-b432-4d70-8704-bbd2bd637691" containerName="extract-utilities" Sep 30 14:22:34 crc kubenswrapper[4840]: I0930 14:22:34.782989 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e5fc6b6-b432-4d70-8704-bbd2bd637691" containerName="extract-utilities" Sep 30 14:22:34 crc kubenswrapper[4840]: E0930 14:22:34.782999 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Sep 30 14:22:34 crc kubenswrapper[4840]: I0930 14:22:34.783006 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Sep 30 14:22:34 crc kubenswrapper[4840]: I0930 14:22:34.783181 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Sep 30 14:22:34 crc kubenswrapper[4840]: I0930 14:22:34.783195 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e5fc6b6-b432-4d70-8704-bbd2bd637691" containerName="registry-server" Sep 30 14:22:34 crc kubenswrapper[4840]: I0930 14:22:34.783827 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xvhf8" Sep 30 14:22:34 crc kubenswrapper[4840]: I0930 14:22:34.793473 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9zm9z" Sep 30 14:22:34 crc kubenswrapper[4840]: I0930 14:22:34.793718 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 14:22:34 crc kubenswrapper[4840]: I0930 14:22:34.793840 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 14:22:34 crc kubenswrapper[4840]: I0930 14:22:34.794704 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 14:22:34 crc kubenswrapper[4840]: I0930 14:22:34.810187 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xvhf8"] Sep 30 14:22:34 crc kubenswrapper[4840]: I0930 14:22:34.938525 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckj5p\" (UniqueName: \"kubernetes.io/projected/b3a0bf4c-6d14-49c4-9a98-3506318ee798-kube-api-access-ckj5p\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xvhf8\" (UID: \"b3a0bf4c-6d14-49c4-9a98-3506318ee798\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xvhf8" Sep 30 14:22:34 crc kubenswrapper[4840]: I0930 14:22:34.938606 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3a0bf4c-6d14-49c4-9a98-3506318ee798-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xvhf8\" (UID: \"b3a0bf4c-6d14-49c4-9a98-3506318ee798\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xvhf8" Sep 30 14:22:34 crc kubenswrapper[4840]: I0930 14:22:34.938653 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3a0bf4c-6d14-49c4-9a98-3506318ee798-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xvhf8\" (UID: \"b3a0bf4c-6d14-49c4-9a98-3506318ee798\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xvhf8" Sep 30 14:22:35 crc kubenswrapper[4840]: I0930 14:22:35.032342 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-d231-account-create-2pgzv"] Sep 30 14:22:35 crc kubenswrapper[4840]: I0930 14:22:35.040211 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckj5p\" (UniqueName: \"kubernetes.io/projected/b3a0bf4c-6d14-49c4-9a98-3506318ee798-kube-api-access-ckj5p\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xvhf8\" (UID: \"b3a0bf4c-6d14-49c4-9a98-3506318ee798\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xvhf8" Sep 30 14:22:35 crc kubenswrapper[4840]: I0930 14:22:35.040268 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3a0bf4c-6d14-49c4-9a98-3506318ee798-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xvhf8\" (UID: \"b3a0bf4c-6d14-49c4-9a98-3506318ee798\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xvhf8" Sep 30 14:22:35 crc kubenswrapper[4840]: I0930 14:22:35.040317 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3a0bf4c-6d14-49c4-9a98-3506318ee798-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xvhf8\" (UID: \"b3a0bf4c-6d14-49c4-9a98-3506318ee798\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xvhf8" Sep 30 14:22:35 crc kubenswrapper[4840]: I0930 14:22:35.042674 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-d231-account-create-2pgzv"] Sep 30 14:22:35 crc kubenswrapper[4840]: I0930 14:22:35.044526 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3a0bf4c-6d14-49c4-9a98-3506318ee798-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xvhf8\" (UID: \"b3a0bf4c-6d14-49c4-9a98-3506318ee798\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xvhf8" Sep 30 14:22:35 crc kubenswrapper[4840]: I0930 14:22:35.046704 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3a0bf4c-6d14-49c4-9a98-3506318ee798-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xvhf8\" (UID: \"b3a0bf4c-6d14-49c4-9a98-3506318ee798\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xvhf8" Sep 30 14:22:35 crc kubenswrapper[4840]: I0930 14:22:35.059614 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckj5p\" (UniqueName: \"kubernetes.io/projected/b3a0bf4c-6d14-49c4-9a98-3506318ee798-kube-api-access-ckj5p\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xvhf8\" (UID: \"b3a0bf4c-6d14-49c4-9a98-3506318ee798\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xvhf8" Sep 30 14:22:35 crc kubenswrapper[4840]: I0930 14:22:35.103362 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xvhf8" Sep 30 14:22:35 crc kubenswrapper[4840]: I0930 14:22:35.587977 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xvhf8"] Sep 30 14:22:35 crc kubenswrapper[4840]: I0930 14:22:35.593931 4840 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 14:22:35 crc kubenswrapper[4840]: I0930 14:22:35.721594 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xvhf8" event={"ID":"b3a0bf4c-6d14-49c4-9a98-3506318ee798","Type":"ContainerStarted","Data":"f0cd98b9fb6d8938ee3f3c7678fa9ab49c1ad0871ab90c8c34ce7f61d617c630"} Sep 30 14:22:36 crc kubenswrapper[4840]: I0930 14:22:36.138324 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa569ef4-7a5f-42f2-9551-b7e491f0792a" path="/var/lib/kubelet/pods/fa569ef4-7a5f-42f2-9551-b7e491f0792a/volumes" Sep 30 14:22:36 crc kubenswrapper[4840]: I0930 14:22:36.729934 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xvhf8" event={"ID":"b3a0bf4c-6d14-49c4-9a98-3506318ee798","Type":"ContainerStarted","Data":"f0045428d7e8bbc1a65f454f83fd4b5edbd159003a4de0ca0e4f8006d2ded873"} Sep 30 14:22:36 crc kubenswrapper[4840]: I0930 14:22:36.752935 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xvhf8" podStartSLOduration=2.158212949 podStartE2EDuration="2.752915985s" podCreationTimestamp="2025-09-30 14:22:34 +0000 UTC" firstStartedPulling="2025-09-30 14:22:35.593340472 +0000 UTC m=+1584.222426935" lastFinishedPulling="2025-09-30 14:22:36.188043538 +0000 UTC m=+1584.817129971" observedRunningTime="2025-09-30 14:22:36.746167731 +0000 UTC m=+1585.375254174" watchObservedRunningTime="2025-09-30 14:22:36.752915985 +0000 UTC m=+1585.382002418" Sep 30 14:22:37 crc kubenswrapper[4840]: I0930 14:22:37.036615 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-dc26-account-create-69bjd"] Sep 30 14:22:37 crc kubenswrapper[4840]: I0930 14:22:37.060359 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-aaa0-account-create-fsw9v"] Sep 30 14:22:37 crc kubenswrapper[4840]: I0930 14:22:37.068306 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-dc26-account-create-69bjd"] Sep 30 14:22:37 crc kubenswrapper[4840]: I0930 14:22:37.075944 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-aaa0-account-create-fsw9v"] Sep 30 14:22:38 crc kubenswrapper[4840]: I0930 14:22:38.025843 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-xvmrt"] Sep 30 14:22:38 crc kubenswrapper[4840]: I0930 14:22:38.035912 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-xvmrt"] Sep 30 14:22:38 crc kubenswrapper[4840]: I0930 14:22:38.127520 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bafeafc-d531-4c3f-8705-65ce26043333" path="/var/lib/kubelet/pods/3bafeafc-d531-4c3f-8705-65ce26043333/volumes" Sep 30 14:22:38 crc kubenswrapper[4840]: I0930 14:22:38.128107 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51a8e57c-0cbf-4072-abb3-696b64c00adb" path="/var/lib/kubelet/pods/51a8e57c-0cbf-4072-abb3-696b64c00adb/volumes" Sep 30 14:22:38 crc kubenswrapper[4840]: I0930 14:22:38.128679 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5749a5e-2f96-49d9-b1c9-d4b8e231b903" path="/var/lib/kubelet/pods/b5749a5e-2f96-49d9-b1c9-d4b8e231b903/volumes" Sep 30 14:23:08 crc kubenswrapper[4840]: I0930 14:23:08.040887 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-nplct"] Sep 30 14:23:08 crc kubenswrapper[4840]: I0930 14:23:08.048876 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-nplct"] Sep 30 14:23:08 crc kubenswrapper[4840]: I0930 14:23:08.127882 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85509efb-abc4-4649-a320-7b9de5487180" path="/var/lib/kubelet/pods/85509efb-abc4-4649-a320-7b9de5487180/volumes" Sep 30 14:23:15 crc kubenswrapper[4840]: I0930 14:23:15.034502 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-5gzc5"] Sep 30 14:23:15 crc kubenswrapper[4840]: I0930 14:23:15.050415 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-5gzc5"] Sep 30 14:23:16 crc kubenswrapper[4840]: I0930 14:23:16.127147 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f907c87-9649-4dd2-a419-97055feb4d18" path="/var/lib/kubelet/pods/0f907c87-9649-4dd2-a419-97055feb4d18/volumes" Sep 30 14:23:21 crc kubenswrapper[4840]: I0930 14:23:21.872123 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:23:21 crc kubenswrapper[4840]: I0930 14:23:21.872799 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:23:27 crc kubenswrapper[4840]: I0930 14:23:27.033157 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-sp7q5"] Sep 30 14:23:27 crc kubenswrapper[4840]: I0930 14:23:27.045467 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-9hcwn"] Sep 30 14:23:27 crc kubenswrapper[4840]: I0930 14:23:27.060133 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-9hcwn"] Sep 30 14:23:27 crc kubenswrapper[4840]: I0930 14:23:27.067540 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-sp7q5"] Sep 30 14:23:28 crc kubenswrapper[4840]: I0930 14:23:28.126326 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c465939-fcbe-44b2-b7ae-a24b5a581dbc" path="/var/lib/kubelet/pods/2c465939-fcbe-44b2-b7ae-a24b5a581dbc/volumes" Sep 30 14:23:28 crc kubenswrapper[4840]: I0930 14:23:28.127456 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad748b83-4edd-4545-9386-032185c531c5" path="/var/lib/kubelet/pods/ad748b83-4edd-4545-9386-032185c531c5/volumes" Sep 30 14:23:30 crc kubenswrapper[4840]: I0930 14:23:30.026820 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-9vd76"] Sep 30 14:23:30 crc kubenswrapper[4840]: I0930 14:23:30.033895 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-9vd76"] Sep 30 14:23:30 crc kubenswrapper[4840]: I0930 14:23:30.127081 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="242d13bb-6113-421b-83d1-fcccf91a9c68" path="/var/lib/kubelet/pods/242d13bb-6113-421b-83d1-fcccf91a9c68/volumes" Sep 30 14:23:34 crc kubenswrapper[4840]: I0930 14:23:34.066102 4840 scope.go:117] "RemoveContainer" containerID="2f4453a841d92d974512f9df0a23b410b7163fc5fd3797f2b601df44aa7d14fb" Sep 30 14:23:34 crc kubenswrapper[4840]: I0930 14:23:34.106665 4840 scope.go:117] "RemoveContainer" containerID="d3c6267fe4fb96f777e003f3fdc08de753fce77bccf131fe0e1b56392dc857cf" Sep 30 14:23:34 crc kubenswrapper[4840]: I0930 14:23:34.135470 4840 scope.go:117] "RemoveContainer" containerID="f4c754f0714009dfde0d74882248a19e96c17147f4c8a912359797cef63027ca" Sep 30 14:23:34 crc kubenswrapper[4840]: I0930 14:23:34.202716 4840 scope.go:117] "RemoveContainer" containerID="58f6da62f61ae533f5086dfb033f031419373cb17ddf4e1ee41d20d765c50f2c" Sep 30 14:23:34 crc kubenswrapper[4840]: I0930 14:23:34.242718 4840 scope.go:117] "RemoveContainer" containerID="0dea90648d1818fd381540b7490dba9e731519612a3256f4d1cda5b6be1ce6fd" Sep 30 14:23:34 crc kubenswrapper[4840]: I0930 14:23:34.287402 4840 scope.go:117] "RemoveContainer" containerID="25f1bac61cdc7bb81eac30ec24cc17df033b36b1740140f97751b58d10a731b7" Sep 30 14:23:34 crc kubenswrapper[4840]: I0930 14:23:34.329314 4840 scope.go:117] "RemoveContainer" containerID="8391b03af598c7e9580a0e3ab6d61f6d82ad2ca29aee7293a25f1fb098fad220" Sep 30 14:23:34 crc kubenswrapper[4840]: I0930 14:23:34.354646 4840 scope.go:117] "RemoveContainer" containerID="22169f9c22fcb2a49c956e27c3ad963f3388f8869b8a6961e0f3920c2c4784b2" Sep 30 14:23:34 crc kubenswrapper[4840]: I0930 14:23:34.376202 4840 scope.go:117] "RemoveContainer" containerID="55c397fd11aa82dcbaa291c03f23c251a2c0e9edbfb49550400ccd57f18fe70e" Sep 30 14:23:34 crc kubenswrapper[4840]: I0930 14:23:34.398569 4840 scope.go:117] "RemoveContainer" containerID="7b37e879a3a49adb8414c595ca1aea847737f1f8f450957b4c08d0a2f64813e5" Sep 30 14:23:50 crc kubenswrapper[4840]: I0930 14:23:50.370246 4840 generic.go:334] "Generic (PLEG): container finished" podID="b3a0bf4c-6d14-49c4-9a98-3506318ee798" containerID="f0045428d7e8bbc1a65f454f83fd4b5edbd159003a4de0ca0e4f8006d2ded873" exitCode=0 Sep 30 14:23:50 crc kubenswrapper[4840]: I0930 14:23:50.370328 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xvhf8" event={"ID":"b3a0bf4c-6d14-49c4-9a98-3506318ee798","Type":"ContainerDied","Data":"f0045428d7e8bbc1a65f454f83fd4b5edbd159003a4de0ca0e4f8006d2ded873"} Sep 30 14:23:51 crc kubenswrapper[4840]: I0930 14:23:51.752256 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xvhf8" Sep 30 14:23:51 crc kubenswrapper[4840]: I0930 14:23:51.842662 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3a0bf4c-6d14-49c4-9a98-3506318ee798-ssh-key\") pod \"b3a0bf4c-6d14-49c4-9a98-3506318ee798\" (UID: \"b3a0bf4c-6d14-49c4-9a98-3506318ee798\") " Sep 30 14:23:51 crc kubenswrapper[4840]: I0930 14:23:51.842812 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3a0bf4c-6d14-49c4-9a98-3506318ee798-inventory\") pod \"b3a0bf4c-6d14-49c4-9a98-3506318ee798\" (UID: \"b3a0bf4c-6d14-49c4-9a98-3506318ee798\") " Sep 30 14:23:51 crc kubenswrapper[4840]: I0930 14:23:51.842858 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ckj5p\" (UniqueName: \"kubernetes.io/projected/b3a0bf4c-6d14-49c4-9a98-3506318ee798-kube-api-access-ckj5p\") pod \"b3a0bf4c-6d14-49c4-9a98-3506318ee798\" (UID: \"b3a0bf4c-6d14-49c4-9a98-3506318ee798\") " Sep 30 14:23:51 crc kubenswrapper[4840]: I0930 14:23:51.849768 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3a0bf4c-6d14-49c4-9a98-3506318ee798-kube-api-access-ckj5p" (OuterVolumeSpecName: "kube-api-access-ckj5p") pod "b3a0bf4c-6d14-49c4-9a98-3506318ee798" (UID: "b3a0bf4c-6d14-49c4-9a98-3506318ee798"). InnerVolumeSpecName "kube-api-access-ckj5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:23:51 crc kubenswrapper[4840]: I0930 14:23:51.871780 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:23:51 crc kubenswrapper[4840]: I0930 14:23:51.871853 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:23:51 crc kubenswrapper[4840]: I0930 14:23:51.872330 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3a0bf4c-6d14-49c4-9a98-3506318ee798-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b3a0bf4c-6d14-49c4-9a98-3506318ee798" (UID: "b3a0bf4c-6d14-49c4-9a98-3506318ee798"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:23:51 crc kubenswrapper[4840]: I0930 14:23:51.872680 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3a0bf4c-6d14-49c4-9a98-3506318ee798-inventory" (OuterVolumeSpecName: "inventory") pod "b3a0bf4c-6d14-49c4-9a98-3506318ee798" (UID: "b3a0bf4c-6d14-49c4-9a98-3506318ee798"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:23:51 crc kubenswrapper[4840]: I0930 14:23:51.945408 4840 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3a0bf4c-6d14-49c4-9a98-3506318ee798-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:23:51 crc kubenswrapper[4840]: I0930 14:23:51.945438 4840 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3a0bf4c-6d14-49c4-9a98-3506318ee798-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 14:23:51 crc kubenswrapper[4840]: I0930 14:23:51.945450 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ckj5p\" (UniqueName: \"kubernetes.io/projected/b3a0bf4c-6d14-49c4-9a98-3506318ee798-kube-api-access-ckj5p\") on node \"crc\" DevicePath \"\"" Sep 30 14:23:52 crc kubenswrapper[4840]: I0930 14:23:52.391783 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xvhf8" event={"ID":"b3a0bf4c-6d14-49c4-9a98-3506318ee798","Type":"ContainerDied","Data":"f0cd98b9fb6d8938ee3f3c7678fa9ab49c1ad0871ab90c8c34ce7f61d617c630"} Sep 30 14:23:52 crc kubenswrapper[4840]: I0930 14:23:52.391842 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f0cd98b9fb6d8938ee3f3c7678fa9ab49c1ad0871ab90c8c34ce7f61d617c630" Sep 30 14:23:52 crc kubenswrapper[4840]: I0930 14:23:52.391884 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xvhf8" Sep 30 14:23:52 crc kubenswrapper[4840]: I0930 14:23:52.488979 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wmtv5"] Sep 30 14:23:52 crc kubenswrapper[4840]: E0930 14:23:52.489720 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3a0bf4c-6d14-49c4-9a98-3506318ee798" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Sep 30 14:23:52 crc kubenswrapper[4840]: I0930 14:23:52.489933 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3a0bf4c-6d14-49c4-9a98-3506318ee798" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Sep 30 14:23:52 crc kubenswrapper[4840]: I0930 14:23:52.490188 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3a0bf4c-6d14-49c4-9a98-3506318ee798" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Sep 30 14:23:52 crc kubenswrapper[4840]: I0930 14:23:52.490926 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wmtv5" Sep 30 14:23:52 crc kubenswrapper[4840]: I0930 14:23:52.494662 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 14:23:52 crc kubenswrapper[4840]: I0930 14:23:52.494662 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 14:23:52 crc kubenswrapper[4840]: I0930 14:23:52.494753 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9zm9z" Sep 30 14:23:52 crc kubenswrapper[4840]: I0930 14:23:52.494908 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 14:23:52 crc kubenswrapper[4840]: I0930 14:23:52.505207 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wmtv5"] Sep 30 14:23:52 crc kubenswrapper[4840]: I0930 14:23:52.556920 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbr22\" (UniqueName: \"kubernetes.io/projected/43ad4b75-0f40-4adc-8bda-67ed967051f3-kube-api-access-dbr22\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wmtv5\" (UID: \"43ad4b75-0f40-4adc-8bda-67ed967051f3\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wmtv5" Sep 30 14:23:52 crc kubenswrapper[4840]: I0930 14:23:52.557196 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43ad4b75-0f40-4adc-8bda-67ed967051f3-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wmtv5\" (UID: \"43ad4b75-0f40-4adc-8bda-67ed967051f3\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wmtv5" Sep 30 14:23:52 crc kubenswrapper[4840]: I0930 14:23:52.557347 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43ad4b75-0f40-4adc-8bda-67ed967051f3-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wmtv5\" (UID: \"43ad4b75-0f40-4adc-8bda-67ed967051f3\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wmtv5" Sep 30 14:23:52 crc kubenswrapper[4840]: I0930 14:23:52.659123 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43ad4b75-0f40-4adc-8bda-67ed967051f3-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wmtv5\" (UID: \"43ad4b75-0f40-4adc-8bda-67ed967051f3\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wmtv5" Sep 30 14:23:52 crc kubenswrapper[4840]: I0930 14:23:52.659871 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbr22\" (UniqueName: \"kubernetes.io/projected/43ad4b75-0f40-4adc-8bda-67ed967051f3-kube-api-access-dbr22\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wmtv5\" (UID: \"43ad4b75-0f40-4adc-8bda-67ed967051f3\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wmtv5" Sep 30 14:23:52 crc kubenswrapper[4840]: I0930 14:23:52.659997 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43ad4b75-0f40-4adc-8bda-67ed967051f3-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wmtv5\" (UID: \"43ad4b75-0f40-4adc-8bda-67ed967051f3\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wmtv5" Sep 30 14:23:52 crc kubenswrapper[4840]: I0930 14:23:52.664121 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43ad4b75-0f40-4adc-8bda-67ed967051f3-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wmtv5\" (UID: \"43ad4b75-0f40-4adc-8bda-67ed967051f3\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wmtv5" Sep 30 14:23:52 crc kubenswrapper[4840]: I0930 14:23:52.664239 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43ad4b75-0f40-4adc-8bda-67ed967051f3-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wmtv5\" (UID: \"43ad4b75-0f40-4adc-8bda-67ed967051f3\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wmtv5" Sep 30 14:23:52 crc kubenswrapper[4840]: I0930 14:23:52.688492 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbr22\" (UniqueName: \"kubernetes.io/projected/43ad4b75-0f40-4adc-8bda-67ed967051f3-kube-api-access-dbr22\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wmtv5\" (UID: \"43ad4b75-0f40-4adc-8bda-67ed967051f3\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wmtv5" Sep 30 14:23:52 crc kubenswrapper[4840]: I0930 14:23:52.815168 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wmtv5" Sep 30 14:23:53 crc kubenswrapper[4840]: I0930 14:23:53.144631 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wmtv5"] Sep 30 14:23:53 crc kubenswrapper[4840]: I0930 14:23:53.402657 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wmtv5" event={"ID":"43ad4b75-0f40-4adc-8bda-67ed967051f3","Type":"ContainerStarted","Data":"b564bb9a37e5f2d36f5a520862684940c6ee114a639028def260559b801886a2"} Sep 30 14:23:54 crc kubenswrapper[4840]: I0930 14:23:54.419622 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wmtv5" event={"ID":"43ad4b75-0f40-4adc-8bda-67ed967051f3","Type":"ContainerStarted","Data":"378c6028cf5d4fa4f8c68f62079bbfe7b1b4e23af2bf00b5e333ea0c896949e3"} Sep 30 14:23:54 crc kubenswrapper[4840]: I0930 14:23:54.452365 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wmtv5" podStartSLOduration=1.996438218 podStartE2EDuration="2.452341305s" podCreationTimestamp="2025-09-30 14:23:52 +0000 UTC" firstStartedPulling="2025-09-30 14:23:53.146847156 +0000 UTC m=+1661.775933579" lastFinishedPulling="2025-09-30 14:23:53.602750223 +0000 UTC m=+1662.231836666" observedRunningTime="2025-09-30 14:23:54.441230141 +0000 UTC m=+1663.070316594" watchObservedRunningTime="2025-09-30 14:23:54.452341305 +0000 UTC m=+1663.081427738" Sep 30 14:23:57 crc kubenswrapper[4840]: I0930 14:23:57.361925 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-95cm9"] Sep 30 14:23:57 crc kubenswrapper[4840]: I0930 14:23:57.365305 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-95cm9" Sep 30 14:23:57 crc kubenswrapper[4840]: I0930 14:23:57.376635 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-95cm9"] Sep 30 14:23:57 crc kubenswrapper[4840]: I0930 14:23:57.444531 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee66f4bd-fee1-482e-9eb2-f09f4203084e-catalog-content\") pod \"redhat-marketplace-95cm9\" (UID: \"ee66f4bd-fee1-482e-9eb2-f09f4203084e\") " pod="openshift-marketplace/redhat-marketplace-95cm9" Sep 30 14:23:57 crc kubenswrapper[4840]: I0930 14:23:57.444627 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2sxk\" (UniqueName: \"kubernetes.io/projected/ee66f4bd-fee1-482e-9eb2-f09f4203084e-kube-api-access-q2sxk\") pod \"redhat-marketplace-95cm9\" (UID: \"ee66f4bd-fee1-482e-9eb2-f09f4203084e\") " pod="openshift-marketplace/redhat-marketplace-95cm9" Sep 30 14:23:57 crc kubenswrapper[4840]: I0930 14:23:57.444661 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee66f4bd-fee1-482e-9eb2-f09f4203084e-utilities\") pod \"redhat-marketplace-95cm9\" (UID: \"ee66f4bd-fee1-482e-9eb2-f09f4203084e\") " pod="openshift-marketplace/redhat-marketplace-95cm9" Sep 30 14:23:57 crc kubenswrapper[4840]: I0930 14:23:57.546492 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee66f4bd-fee1-482e-9eb2-f09f4203084e-catalog-content\") pod \"redhat-marketplace-95cm9\" (UID: \"ee66f4bd-fee1-482e-9eb2-f09f4203084e\") " pod="openshift-marketplace/redhat-marketplace-95cm9" Sep 30 14:23:57 crc kubenswrapper[4840]: I0930 14:23:57.546591 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2sxk\" (UniqueName: \"kubernetes.io/projected/ee66f4bd-fee1-482e-9eb2-f09f4203084e-kube-api-access-q2sxk\") pod \"redhat-marketplace-95cm9\" (UID: \"ee66f4bd-fee1-482e-9eb2-f09f4203084e\") " pod="openshift-marketplace/redhat-marketplace-95cm9" Sep 30 14:23:57 crc kubenswrapper[4840]: I0930 14:23:57.546619 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee66f4bd-fee1-482e-9eb2-f09f4203084e-utilities\") pod \"redhat-marketplace-95cm9\" (UID: \"ee66f4bd-fee1-482e-9eb2-f09f4203084e\") " pod="openshift-marketplace/redhat-marketplace-95cm9" Sep 30 14:23:57 crc kubenswrapper[4840]: I0930 14:23:57.547030 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee66f4bd-fee1-482e-9eb2-f09f4203084e-utilities\") pod \"redhat-marketplace-95cm9\" (UID: \"ee66f4bd-fee1-482e-9eb2-f09f4203084e\") " pod="openshift-marketplace/redhat-marketplace-95cm9" Sep 30 14:23:57 crc kubenswrapper[4840]: I0930 14:23:57.547234 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee66f4bd-fee1-482e-9eb2-f09f4203084e-catalog-content\") pod \"redhat-marketplace-95cm9\" (UID: \"ee66f4bd-fee1-482e-9eb2-f09f4203084e\") " pod="openshift-marketplace/redhat-marketplace-95cm9" Sep 30 14:23:57 crc kubenswrapper[4840]: I0930 14:23:57.568487 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2sxk\" (UniqueName: \"kubernetes.io/projected/ee66f4bd-fee1-482e-9eb2-f09f4203084e-kube-api-access-q2sxk\") pod \"redhat-marketplace-95cm9\" (UID: \"ee66f4bd-fee1-482e-9eb2-f09f4203084e\") " pod="openshift-marketplace/redhat-marketplace-95cm9" Sep 30 14:23:57 crc kubenswrapper[4840]: I0930 14:23:57.701277 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-95cm9" Sep 30 14:23:58 crc kubenswrapper[4840]: W0930 14:23:58.165609 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee66f4bd_fee1_482e_9eb2_f09f4203084e.slice/crio-619c1184472f134d615df52bebee6de76f26ea095a66a5e1523a0dd3c362f3b1 WatchSource:0}: Error finding container 619c1184472f134d615df52bebee6de76f26ea095a66a5e1523a0dd3c362f3b1: Status 404 returned error can't find the container with id 619c1184472f134d615df52bebee6de76f26ea095a66a5e1523a0dd3c362f3b1 Sep 30 14:23:58 crc kubenswrapper[4840]: I0930 14:23:58.171368 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-95cm9"] Sep 30 14:23:58 crc kubenswrapper[4840]: I0930 14:23:58.452067 4840 generic.go:334] "Generic (PLEG): container finished" podID="43ad4b75-0f40-4adc-8bda-67ed967051f3" containerID="378c6028cf5d4fa4f8c68f62079bbfe7b1b4e23af2bf00b5e333ea0c896949e3" exitCode=0 Sep 30 14:23:58 crc kubenswrapper[4840]: I0930 14:23:58.452162 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wmtv5" event={"ID":"43ad4b75-0f40-4adc-8bda-67ed967051f3","Type":"ContainerDied","Data":"378c6028cf5d4fa4f8c68f62079bbfe7b1b4e23af2bf00b5e333ea0c896949e3"} Sep 30 14:23:58 crc kubenswrapper[4840]: I0930 14:23:58.454650 4840 generic.go:334] "Generic (PLEG): container finished" podID="ee66f4bd-fee1-482e-9eb2-f09f4203084e" containerID="7dbe9c71791c8ed39428e0575e46a322f5f41d5dca98407b3a7a33822b5d2ca9" exitCode=0 Sep 30 14:23:58 crc kubenswrapper[4840]: I0930 14:23:58.454696 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-95cm9" event={"ID":"ee66f4bd-fee1-482e-9eb2-f09f4203084e","Type":"ContainerDied","Data":"7dbe9c71791c8ed39428e0575e46a322f5f41d5dca98407b3a7a33822b5d2ca9"} Sep 30 14:23:58 crc kubenswrapper[4840]: I0930 14:23:58.454721 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-95cm9" event={"ID":"ee66f4bd-fee1-482e-9eb2-f09f4203084e","Type":"ContainerStarted","Data":"619c1184472f134d615df52bebee6de76f26ea095a66a5e1523a0dd3c362f3b1"} Sep 30 14:23:59 crc kubenswrapper[4840]: I0930 14:23:59.959842 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wmtv5" Sep 30 14:24:00 crc kubenswrapper[4840]: I0930 14:24:00.097220 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43ad4b75-0f40-4adc-8bda-67ed967051f3-inventory\") pod \"43ad4b75-0f40-4adc-8bda-67ed967051f3\" (UID: \"43ad4b75-0f40-4adc-8bda-67ed967051f3\") " Sep 30 14:24:00 crc kubenswrapper[4840]: I0930 14:24:00.097262 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43ad4b75-0f40-4adc-8bda-67ed967051f3-ssh-key\") pod \"43ad4b75-0f40-4adc-8bda-67ed967051f3\" (UID: \"43ad4b75-0f40-4adc-8bda-67ed967051f3\") " Sep 30 14:24:00 crc kubenswrapper[4840]: I0930 14:24:00.097490 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbr22\" (UniqueName: \"kubernetes.io/projected/43ad4b75-0f40-4adc-8bda-67ed967051f3-kube-api-access-dbr22\") pod \"43ad4b75-0f40-4adc-8bda-67ed967051f3\" (UID: \"43ad4b75-0f40-4adc-8bda-67ed967051f3\") " Sep 30 14:24:00 crc kubenswrapper[4840]: I0930 14:24:00.106748 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43ad4b75-0f40-4adc-8bda-67ed967051f3-kube-api-access-dbr22" (OuterVolumeSpecName: "kube-api-access-dbr22") pod "43ad4b75-0f40-4adc-8bda-67ed967051f3" (UID: "43ad4b75-0f40-4adc-8bda-67ed967051f3"). InnerVolumeSpecName "kube-api-access-dbr22". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:24:00 crc kubenswrapper[4840]: I0930 14:24:00.129188 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43ad4b75-0f40-4adc-8bda-67ed967051f3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "43ad4b75-0f40-4adc-8bda-67ed967051f3" (UID: "43ad4b75-0f40-4adc-8bda-67ed967051f3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:24:00 crc kubenswrapper[4840]: I0930 14:24:00.134489 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43ad4b75-0f40-4adc-8bda-67ed967051f3-inventory" (OuterVolumeSpecName: "inventory") pod "43ad4b75-0f40-4adc-8bda-67ed967051f3" (UID: "43ad4b75-0f40-4adc-8bda-67ed967051f3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:24:00 crc kubenswrapper[4840]: I0930 14:24:00.199481 4840 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43ad4b75-0f40-4adc-8bda-67ed967051f3-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 14:24:00 crc kubenswrapper[4840]: I0930 14:24:00.199510 4840 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43ad4b75-0f40-4adc-8bda-67ed967051f3-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:24:00 crc kubenswrapper[4840]: I0930 14:24:00.199520 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbr22\" (UniqueName: \"kubernetes.io/projected/43ad4b75-0f40-4adc-8bda-67ed967051f3-kube-api-access-dbr22\") on node \"crc\" DevicePath \"\"" Sep 30 14:24:00 crc kubenswrapper[4840]: I0930 14:24:00.474793 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wmtv5" event={"ID":"43ad4b75-0f40-4adc-8bda-67ed967051f3","Type":"ContainerDied","Data":"b564bb9a37e5f2d36f5a520862684940c6ee114a639028def260559b801886a2"} Sep 30 14:24:00 crc kubenswrapper[4840]: I0930 14:24:00.475107 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b564bb9a37e5f2d36f5a520862684940c6ee114a639028def260559b801886a2" Sep 30 14:24:00 crc kubenswrapper[4840]: I0930 14:24:00.474863 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wmtv5" Sep 30 14:24:00 crc kubenswrapper[4840]: I0930 14:24:00.476844 4840 generic.go:334] "Generic (PLEG): container finished" podID="ee66f4bd-fee1-482e-9eb2-f09f4203084e" containerID="6daa00aa433eda13a857918fe56918449116c371e6fb02dcd783e49a2ca4a8c4" exitCode=0 Sep 30 14:24:00 crc kubenswrapper[4840]: I0930 14:24:00.476869 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-95cm9" event={"ID":"ee66f4bd-fee1-482e-9eb2-f09f4203084e","Type":"ContainerDied","Data":"6daa00aa433eda13a857918fe56918449116c371e6fb02dcd783e49a2ca4a8c4"} Sep 30 14:24:00 crc kubenswrapper[4840]: I0930 14:24:00.561099 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-hw58j"] Sep 30 14:24:00 crc kubenswrapper[4840]: E0930 14:24:00.561444 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43ad4b75-0f40-4adc-8bda-67ed967051f3" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Sep 30 14:24:00 crc kubenswrapper[4840]: I0930 14:24:00.561463 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="43ad4b75-0f40-4adc-8bda-67ed967051f3" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Sep 30 14:24:00 crc kubenswrapper[4840]: I0930 14:24:00.561667 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="43ad4b75-0f40-4adc-8bda-67ed967051f3" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Sep 30 14:24:00 crc kubenswrapper[4840]: I0930 14:24:00.562312 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hw58j" Sep 30 14:24:00 crc kubenswrapper[4840]: I0930 14:24:00.569029 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9zm9z" Sep 30 14:24:00 crc kubenswrapper[4840]: I0930 14:24:00.569044 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 14:24:00 crc kubenswrapper[4840]: I0930 14:24:00.569150 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 14:24:00 crc kubenswrapper[4840]: I0930 14:24:00.569353 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 14:24:00 crc kubenswrapper[4840]: I0930 14:24:00.573821 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-hw58j"] Sep 30 14:24:00 crc kubenswrapper[4840]: I0930 14:24:00.708763 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3c378bd2-1562-466e-aafb-7a597487f2c0-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hw58j\" (UID: \"3c378bd2-1562-466e-aafb-7a597487f2c0\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hw58j" Sep 30 14:24:00 crc kubenswrapper[4840]: I0930 14:24:00.708889 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8gn8\" (UniqueName: \"kubernetes.io/projected/3c378bd2-1562-466e-aafb-7a597487f2c0-kube-api-access-c8gn8\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hw58j\" (UID: \"3c378bd2-1562-466e-aafb-7a597487f2c0\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hw58j" Sep 30 14:24:00 crc kubenswrapper[4840]: I0930 14:24:00.708962 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3c378bd2-1562-466e-aafb-7a597487f2c0-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hw58j\" (UID: \"3c378bd2-1562-466e-aafb-7a597487f2c0\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hw58j" Sep 30 14:24:00 crc kubenswrapper[4840]: I0930 14:24:00.811144 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3c378bd2-1562-466e-aafb-7a597487f2c0-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hw58j\" (UID: \"3c378bd2-1562-466e-aafb-7a597487f2c0\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hw58j" Sep 30 14:24:00 crc kubenswrapper[4840]: I0930 14:24:00.811663 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8gn8\" (UniqueName: \"kubernetes.io/projected/3c378bd2-1562-466e-aafb-7a597487f2c0-kube-api-access-c8gn8\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hw58j\" (UID: \"3c378bd2-1562-466e-aafb-7a597487f2c0\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hw58j" Sep 30 14:24:00 crc kubenswrapper[4840]: I0930 14:24:00.811758 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3c378bd2-1562-466e-aafb-7a597487f2c0-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hw58j\" (UID: \"3c378bd2-1562-466e-aafb-7a597487f2c0\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hw58j" Sep 30 14:24:00 crc kubenswrapper[4840]: I0930 14:24:00.818118 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3c378bd2-1562-466e-aafb-7a597487f2c0-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hw58j\" (UID: \"3c378bd2-1562-466e-aafb-7a597487f2c0\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hw58j" Sep 30 14:24:00 crc kubenswrapper[4840]: I0930 14:24:00.818980 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3c378bd2-1562-466e-aafb-7a597487f2c0-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hw58j\" (UID: \"3c378bd2-1562-466e-aafb-7a597487f2c0\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hw58j" Sep 30 14:24:00 crc kubenswrapper[4840]: I0930 14:24:00.842601 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8gn8\" (UniqueName: \"kubernetes.io/projected/3c378bd2-1562-466e-aafb-7a597487f2c0-kube-api-access-c8gn8\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hw58j\" (UID: \"3c378bd2-1562-466e-aafb-7a597487f2c0\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hw58j" Sep 30 14:24:00 crc kubenswrapper[4840]: I0930 14:24:00.883050 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hw58j" Sep 30 14:24:01 crc kubenswrapper[4840]: I0930 14:24:01.387539 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-hw58j"] Sep 30 14:24:01 crc kubenswrapper[4840]: W0930 14:24:01.389634 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c378bd2_1562_466e_aafb_7a597487f2c0.slice/crio-8785f364cf1d9e7efea290d9450dc3caf61c1322604d3e39247290edcae2d3a8 WatchSource:0}: Error finding container 8785f364cf1d9e7efea290d9450dc3caf61c1322604d3e39247290edcae2d3a8: Status 404 returned error can't find the container with id 8785f364cf1d9e7efea290d9450dc3caf61c1322604d3e39247290edcae2d3a8 Sep 30 14:24:01 crc kubenswrapper[4840]: I0930 14:24:01.490530 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-95cm9" event={"ID":"ee66f4bd-fee1-482e-9eb2-f09f4203084e","Type":"ContainerStarted","Data":"807a3c5bc3c468ce47b480d856d103ce39b9e16d55ea1f874c9b8caa7b156337"} Sep 30 14:24:01 crc kubenswrapper[4840]: I0930 14:24:01.493739 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hw58j" event={"ID":"3c378bd2-1562-466e-aafb-7a597487f2c0","Type":"ContainerStarted","Data":"8785f364cf1d9e7efea290d9450dc3caf61c1322604d3e39247290edcae2d3a8"} Sep 30 14:24:01 crc kubenswrapper[4840]: I0930 14:24:01.508960 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-95cm9" podStartSLOduration=1.707380776 podStartE2EDuration="4.50893856s" podCreationTimestamp="2025-09-30 14:23:57 +0000 UTC" firstStartedPulling="2025-09-30 14:23:58.456815984 +0000 UTC m=+1667.085902407" lastFinishedPulling="2025-09-30 14:24:01.258373758 +0000 UTC m=+1669.887460191" observedRunningTime="2025-09-30 14:24:01.506479607 +0000 UTC m=+1670.135566030" watchObservedRunningTime="2025-09-30 14:24:01.50893856 +0000 UTC m=+1670.138024993" Sep 30 14:24:02 crc kubenswrapper[4840]: I0930 14:24:02.506713 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hw58j" event={"ID":"3c378bd2-1562-466e-aafb-7a597487f2c0","Type":"ContainerStarted","Data":"dd315464fab31f9c7bda66ccb13fe97b5076326ee0ba307fedd8fcf005472bc3"} Sep 30 14:24:02 crc kubenswrapper[4840]: I0930 14:24:02.525717 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hw58j" podStartSLOduration=2.001776062 podStartE2EDuration="2.52570078s" podCreationTimestamp="2025-09-30 14:24:00 +0000 UTC" firstStartedPulling="2025-09-30 14:24:01.392390527 +0000 UTC m=+1670.021476950" lastFinishedPulling="2025-09-30 14:24:01.916315245 +0000 UTC m=+1670.545401668" observedRunningTime="2025-09-30 14:24:02.522479478 +0000 UTC m=+1671.151565911" watchObservedRunningTime="2025-09-30 14:24:02.52570078 +0000 UTC m=+1671.154787203" Sep 30 14:24:05 crc kubenswrapper[4840]: I0930 14:24:05.052830 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-g7d5k"] Sep 30 14:24:05 crc kubenswrapper[4840]: I0930 14:24:05.065534 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-xgj9n"] Sep 30 14:24:05 crc kubenswrapper[4840]: I0930 14:24:05.075820 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-g7d5k"] Sep 30 14:24:05 crc kubenswrapper[4840]: I0930 14:24:05.084894 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-xgj9n"] Sep 30 14:24:05 crc kubenswrapper[4840]: I0930 14:24:05.093021 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-r8kd7"] Sep 30 14:24:05 crc kubenswrapper[4840]: I0930 14:24:05.100773 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-r8kd7"] Sep 30 14:24:06 crc kubenswrapper[4840]: I0930 14:24:06.131023 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12f02cd8-eb57-4557-b08a-1fc38f1f49fd" path="/var/lib/kubelet/pods/12f02cd8-eb57-4557-b08a-1fc38f1f49fd/volumes" Sep 30 14:24:06 crc kubenswrapper[4840]: I0930 14:24:06.132656 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="252dac07-3c44-4264-a488-cca59b98e516" path="/var/lib/kubelet/pods/252dac07-3c44-4264-a488-cca59b98e516/volumes" Sep 30 14:24:06 crc kubenswrapper[4840]: I0930 14:24:06.133763 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e00317b4-982d-45c8-b5d6-6b69e4d1d0a7" path="/var/lib/kubelet/pods/e00317b4-982d-45c8-b5d6-6b69e4d1d0a7/volumes" Sep 30 14:24:07 crc kubenswrapper[4840]: I0930 14:24:07.701910 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-95cm9" Sep 30 14:24:07 crc kubenswrapper[4840]: I0930 14:24:07.702273 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-95cm9" Sep 30 14:24:07 crc kubenswrapper[4840]: I0930 14:24:07.753776 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-95cm9" Sep 30 14:24:08 crc kubenswrapper[4840]: I0930 14:24:08.607427 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-95cm9" Sep 30 14:24:08 crc kubenswrapper[4840]: I0930 14:24:08.648717 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-95cm9"] Sep 30 14:24:10 crc kubenswrapper[4840]: I0930 14:24:10.575964 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-95cm9" podUID="ee66f4bd-fee1-482e-9eb2-f09f4203084e" containerName="registry-server" containerID="cri-o://807a3c5bc3c468ce47b480d856d103ce39b9e16d55ea1f874c9b8caa7b156337" gracePeriod=2 Sep 30 14:24:11 crc kubenswrapper[4840]: I0930 14:24:11.009685 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-95cm9" Sep 30 14:24:11 crc kubenswrapper[4840]: I0930 14:24:11.093785 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee66f4bd-fee1-482e-9eb2-f09f4203084e-catalog-content\") pod \"ee66f4bd-fee1-482e-9eb2-f09f4203084e\" (UID: \"ee66f4bd-fee1-482e-9eb2-f09f4203084e\") " Sep 30 14:24:11 crc kubenswrapper[4840]: I0930 14:24:11.093847 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2sxk\" (UniqueName: \"kubernetes.io/projected/ee66f4bd-fee1-482e-9eb2-f09f4203084e-kube-api-access-q2sxk\") pod \"ee66f4bd-fee1-482e-9eb2-f09f4203084e\" (UID: \"ee66f4bd-fee1-482e-9eb2-f09f4203084e\") " Sep 30 14:24:11 crc kubenswrapper[4840]: I0930 14:24:11.093900 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee66f4bd-fee1-482e-9eb2-f09f4203084e-utilities\") pod \"ee66f4bd-fee1-482e-9eb2-f09f4203084e\" (UID: \"ee66f4bd-fee1-482e-9eb2-f09f4203084e\") " Sep 30 14:24:11 crc kubenswrapper[4840]: I0930 14:24:11.095197 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee66f4bd-fee1-482e-9eb2-f09f4203084e-utilities" (OuterVolumeSpecName: "utilities") pod "ee66f4bd-fee1-482e-9eb2-f09f4203084e" (UID: "ee66f4bd-fee1-482e-9eb2-f09f4203084e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:24:11 crc kubenswrapper[4840]: I0930 14:24:11.101811 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee66f4bd-fee1-482e-9eb2-f09f4203084e-kube-api-access-q2sxk" (OuterVolumeSpecName: "kube-api-access-q2sxk") pod "ee66f4bd-fee1-482e-9eb2-f09f4203084e" (UID: "ee66f4bd-fee1-482e-9eb2-f09f4203084e"). InnerVolumeSpecName "kube-api-access-q2sxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:24:11 crc kubenswrapper[4840]: I0930 14:24:11.110275 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee66f4bd-fee1-482e-9eb2-f09f4203084e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ee66f4bd-fee1-482e-9eb2-f09f4203084e" (UID: "ee66f4bd-fee1-482e-9eb2-f09f4203084e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:24:11 crc kubenswrapper[4840]: I0930 14:24:11.195521 4840 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee66f4bd-fee1-482e-9eb2-f09f4203084e-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:24:11 crc kubenswrapper[4840]: I0930 14:24:11.195582 4840 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee66f4bd-fee1-482e-9eb2-f09f4203084e-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:24:11 crc kubenswrapper[4840]: I0930 14:24:11.195596 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2sxk\" (UniqueName: \"kubernetes.io/projected/ee66f4bd-fee1-482e-9eb2-f09f4203084e-kube-api-access-q2sxk\") on node \"crc\" DevicePath \"\"" Sep 30 14:24:11 crc kubenswrapper[4840]: I0930 14:24:11.587112 4840 generic.go:334] "Generic (PLEG): container finished" podID="ee66f4bd-fee1-482e-9eb2-f09f4203084e" containerID="807a3c5bc3c468ce47b480d856d103ce39b9e16d55ea1f874c9b8caa7b156337" exitCode=0 Sep 30 14:24:11 crc kubenswrapper[4840]: I0930 14:24:11.587170 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-95cm9" event={"ID":"ee66f4bd-fee1-482e-9eb2-f09f4203084e","Type":"ContainerDied","Data":"807a3c5bc3c468ce47b480d856d103ce39b9e16d55ea1f874c9b8caa7b156337"} Sep 30 14:24:11 crc kubenswrapper[4840]: I0930 14:24:11.587223 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-95cm9" Sep 30 14:24:11 crc kubenswrapper[4840]: I0930 14:24:11.587475 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-95cm9" event={"ID":"ee66f4bd-fee1-482e-9eb2-f09f4203084e","Type":"ContainerDied","Data":"619c1184472f134d615df52bebee6de76f26ea095a66a5e1523a0dd3c362f3b1"} Sep 30 14:24:11 crc kubenswrapper[4840]: I0930 14:24:11.587501 4840 scope.go:117] "RemoveContainer" containerID="807a3c5bc3c468ce47b480d856d103ce39b9e16d55ea1f874c9b8caa7b156337" Sep 30 14:24:11 crc kubenswrapper[4840]: I0930 14:24:11.624456 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-95cm9"] Sep 30 14:24:11 crc kubenswrapper[4840]: I0930 14:24:11.627708 4840 scope.go:117] "RemoveContainer" containerID="6daa00aa433eda13a857918fe56918449116c371e6fb02dcd783e49a2ca4a8c4" Sep 30 14:24:11 crc kubenswrapper[4840]: I0930 14:24:11.634713 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-95cm9"] Sep 30 14:24:11 crc kubenswrapper[4840]: I0930 14:24:11.650054 4840 scope.go:117] "RemoveContainer" containerID="7dbe9c71791c8ed39428e0575e46a322f5f41d5dca98407b3a7a33822b5d2ca9" Sep 30 14:24:11 crc kubenswrapper[4840]: I0930 14:24:11.687866 4840 scope.go:117] "RemoveContainer" containerID="807a3c5bc3c468ce47b480d856d103ce39b9e16d55ea1f874c9b8caa7b156337" Sep 30 14:24:11 crc kubenswrapper[4840]: E0930 14:24:11.688393 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"807a3c5bc3c468ce47b480d856d103ce39b9e16d55ea1f874c9b8caa7b156337\": container with ID starting with 807a3c5bc3c468ce47b480d856d103ce39b9e16d55ea1f874c9b8caa7b156337 not found: ID does not exist" containerID="807a3c5bc3c468ce47b480d856d103ce39b9e16d55ea1f874c9b8caa7b156337" Sep 30 14:24:11 crc kubenswrapper[4840]: I0930 14:24:11.688441 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"807a3c5bc3c468ce47b480d856d103ce39b9e16d55ea1f874c9b8caa7b156337"} err="failed to get container status \"807a3c5bc3c468ce47b480d856d103ce39b9e16d55ea1f874c9b8caa7b156337\": rpc error: code = NotFound desc = could not find container \"807a3c5bc3c468ce47b480d856d103ce39b9e16d55ea1f874c9b8caa7b156337\": container with ID starting with 807a3c5bc3c468ce47b480d856d103ce39b9e16d55ea1f874c9b8caa7b156337 not found: ID does not exist" Sep 30 14:24:11 crc kubenswrapper[4840]: I0930 14:24:11.688470 4840 scope.go:117] "RemoveContainer" containerID="6daa00aa433eda13a857918fe56918449116c371e6fb02dcd783e49a2ca4a8c4" Sep 30 14:24:11 crc kubenswrapper[4840]: E0930 14:24:11.688925 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6daa00aa433eda13a857918fe56918449116c371e6fb02dcd783e49a2ca4a8c4\": container with ID starting with 6daa00aa433eda13a857918fe56918449116c371e6fb02dcd783e49a2ca4a8c4 not found: ID does not exist" containerID="6daa00aa433eda13a857918fe56918449116c371e6fb02dcd783e49a2ca4a8c4" Sep 30 14:24:11 crc kubenswrapper[4840]: I0930 14:24:11.688964 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6daa00aa433eda13a857918fe56918449116c371e6fb02dcd783e49a2ca4a8c4"} err="failed to get container status \"6daa00aa433eda13a857918fe56918449116c371e6fb02dcd783e49a2ca4a8c4\": rpc error: code = NotFound desc = could not find container \"6daa00aa433eda13a857918fe56918449116c371e6fb02dcd783e49a2ca4a8c4\": container with ID starting with 6daa00aa433eda13a857918fe56918449116c371e6fb02dcd783e49a2ca4a8c4 not found: ID does not exist" Sep 30 14:24:11 crc kubenswrapper[4840]: I0930 14:24:11.688994 4840 scope.go:117] "RemoveContainer" containerID="7dbe9c71791c8ed39428e0575e46a322f5f41d5dca98407b3a7a33822b5d2ca9" Sep 30 14:24:11 crc kubenswrapper[4840]: E0930 14:24:11.689282 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7dbe9c71791c8ed39428e0575e46a322f5f41d5dca98407b3a7a33822b5d2ca9\": container with ID starting with 7dbe9c71791c8ed39428e0575e46a322f5f41d5dca98407b3a7a33822b5d2ca9 not found: ID does not exist" containerID="7dbe9c71791c8ed39428e0575e46a322f5f41d5dca98407b3a7a33822b5d2ca9" Sep 30 14:24:11 crc kubenswrapper[4840]: I0930 14:24:11.689314 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dbe9c71791c8ed39428e0575e46a322f5f41d5dca98407b3a7a33822b5d2ca9"} err="failed to get container status \"7dbe9c71791c8ed39428e0575e46a322f5f41d5dca98407b3a7a33822b5d2ca9\": rpc error: code = NotFound desc = could not find container \"7dbe9c71791c8ed39428e0575e46a322f5f41d5dca98407b3a7a33822b5d2ca9\": container with ID starting with 7dbe9c71791c8ed39428e0575e46a322f5f41d5dca98407b3a7a33822b5d2ca9 not found: ID does not exist" Sep 30 14:24:12 crc kubenswrapper[4840]: I0930 14:24:12.126659 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee66f4bd-fee1-482e-9eb2-f09f4203084e" path="/var/lib/kubelet/pods/ee66f4bd-fee1-482e-9eb2-f09f4203084e/volumes" Sep 30 14:24:13 crc kubenswrapper[4840]: I0930 14:24:13.039411 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-6731-account-create-v45sj"] Sep 30 14:24:13 crc kubenswrapper[4840]: I0930 14:24:13.050506 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-60a3-account-create-qjn85"] Sep 30 14:24:13 crc kubenswrapper[4840]: I0930 14:24:13.059975 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-60a3-account-create-qjn85"] Sep 30 14:24:13 crc kubenswrapper[4840]: I0930 14:24:13.066855 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-6731-account-create-v45sj"] Sep 30 14:24:14 crc kubenswrapper[4840]: I0930 14:24:14.127509 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fe3cf3b-8d33-4689-86c9-3331fdcc6fac" path="/var/lib/kubelet/pods/3fe3cf3b-8d33-4689-86c9-3331fdcc6fac/volumes" Sep 30 14:24:14 crc kubenswrapper[4840]: I0930 14:24:14.128049 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c2fd404-193d-4a27-92c9-395b2c0fcc02" path="/var/lib/kubelet/pods/4c2fd404-193d-4a27-92c9-395b2c0fcc02/volumes" Sep 30 14:24:21 crc kubenswrapper[4840]: I0930 14:24:21.030067 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-c6a4-account-create-r4fhv"] Sep 30 14:24:21 crc kubenswrapper[4840]: I0930 14:24:21.036827 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-c6a4-account-create-r4fhv"] Sep 30 14:24:21 crc kubenswrapper[4840]: I0930 14:24:21.871223 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:24:21 crc kubenswrapper[4840]: I0930 14:24:21.871532 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:24:21 crc kubenswrapper[4840]: I0930 14:24:21.871622 4840 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-747gk" Sep 30 14:24:21 crc kubenswrapper[4840]: I0930 14:24:21.872317 4840 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"74b12cc9318ac99161f1b2541861d89b6100ba9a986e5863f53ae26c2a8953e7"} pod="openshift-machine-config-operator/machine-config-daemon-747gk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 14:24:21 crc kubenswrapper[4840]: I0930 14:24:21.872384 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" containerID="cri-o://74b12cc9318ac99161f1b2541861d89b6100ba9a986e5863f53ae26c2a8953e7" gracePeriod=600 Sep 30 14:24:22 crc kubenswrapper[4840]: E0930 14:24:22.001766 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:24:22 crc kubenswrapper[4840]: I0930 14:24:22.128098 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f883b69-fee7-426a-9522-79e1c918e5fb" path="/var/lib/kubelet/pods/6f883b69-fee7-426a-9522-79e1c918e5fb/volumes" Sep 30 14:24:22 crc kubenswrapper[4840]: I0930 14:24:22.693715 4840 generic.go:334] "Generic (PLEG): container finished" podID="10e8b890-7f20-4a36-8e03-898620cf599a" containerID="74b12cc9318ac99161f1b2541861d89b6100ba9a986e5863f53ae26c2a8953e7" exitCode=0 Sep 30 14:24:22 crc kubenswrapper[4840]: I0930 14:24:22.693753 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" event={"ID":"10e8b890-7f20-4a36-8e03-898620cf599a","Type":"ContainerDied","Data":"74b12cc9318ac99161f1b2541861d89b6100ba9a986e5863f53ae26c2a8953e7"} Sep 30 14:24:22 crc kubenswrapper[4840]: I0930 14:24:22.693782 4840 scope.go:117] "RemoveContainer" containerID="b4ec147091f38f805d294a67b19870b478b6bf3d54869eabd43f91f49541a76e" Sep 30 14:24:22 crc kubenswrapper[4840]: I0930 14:24:22.694366 4840 scope.go:117] "RemoveContainer" containerID="74b12cc9318ac99161f1b2541861d89b6100ba9a986e5863f53ae26c2a8953e7" Sep 30 14:24:22 crc kubenswrapper[4840]: E0930 14:24:22.694675 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:24:34 crc kubenswrapper[4840]: I0930 14:24:34.117391 4840 scope.go:117] "RemoveContainer" containerID="74b12cc9318ac99161f1b2541861d89b6100ba9a986e5863f53ae26c2a8953e7" Sep 30 14:24:34 crc kubenswrapper[4840]: E0930 14:24:34.118278 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:24:34 crc kubenswrapper[4840]: I0930 14:24:34.571060 4840 scope.go:117] "RemoveContainer" containerID="333296e5f81bf6eb25692aa0848f4ac2b2c64935f8ca8bb9b5804f26e314562f" Sep 30 14:24:34 crc kubenswrapper[4840]: I0930 14:24:34.611474 4840 scope.go:117] "RemoveContainer" containerID="5a980ce6f702025cb46efebfcd6baf9b79b79afd7f53a1b121cd10497d1b3ad1" Sep 30 14:24:34 crc kubenswrapper[4840]: I0930 14:24:34.644375 4840 scope.go:117] "RemoveContainer" containerID="c1bf996a56aa4ade8ada32f7bee83c4df3bacd4b72b8b401a2c1597ee6724fa5" Sep 30 14:24:34 crc kubenswrapper[4840]: I0930 14:24:34.695344 4840 scope.go:117] "RemoveContainer" containerID="f513d577284ff2ffc61e31d9e5458e15e06f530cc0498bd6d4e92a4378473124" Sep 30 14:24:34 crc kubenswrapper[4840]: I0930 14:24:34.736046 4840 scope.go:117] "RemoveContainer" containerID="455ebb76c7cfb7ea9121ca949e9eae745536c621e9b7c76bc7e8ad9ec7eea647" Sep 30 14:24:34 crc kubenswrapper[4840]: I0930 14:24:34.777938 4840 scope.go:117] "RemoveContainer" containerID="d8633210762a0973646719f652972b15063961d632b50a670d911d219e79321e" Sep 30 14:24:38 crc kubenswrapper[4840]: I0930 14:24:38.840282 4840 generic.go:334] "Generic (PLEG): container finished" podID="3c378bd2-1562-466e-aafb-7a597487f2c0" containerID="dd315464fab31f9c7bda66ccb13fe97b5076326ee0ba307fedd8fcf005472bc3" exitCode=0 Sep 30 14:24:38 crc kubenswrapper[4840]: I0930 14:24:38.840394 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hw58j" event={"ID":"3c378bd2-1562-466e-aafb-7a597487f2c0","Type":"ContainerDied","Data":"dd315464fab31f9c7bda66ccb13fe97b5076326ee0ba307fedd8fcf005472bc3"} Sep 30 14:24:40 crc kubenswrapper[4840]: I0930 14:24:40.257588 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hw58j" Sep 30 14:24:40 crc kubenswrapper[4840]: I0930 14:24:40.423917 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3c378bd2-1562-466e-aafb-7a597487f2c0-ssh-key\") pod \"3c378bd2-1562-466e-aafb-7a597487f2c0\" (UID: \"3c378bd2-1562-466e-aafb-7a597487f2c0\") " Sep 30 14:24:40 crc kubenswrapper[4840]: I0930 14:24:40.423992 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3c378bd2-1562-466e-aafb-7a597487f2c0-inventory\") pod \"3c378bd2-1562-466e-aafb-7a597487f2c0\" (UID: \"3c378bd2-1562-466e-aafb-7a597487f2c0\") " Sep 30 14:24:40 crc kubenswrapper[4840]: I0930 14:24:40.424026 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8gn8\" (UniqueName: \"kubernetes.io/projected/3c378bd2-1562-466e-aafb-7a597487f2c0-kube-api-access-c8gn8\") pod \"3c378bd2-1562-466e-aafb-7a597487f2c0\" (UID: \"3c378bd2-1562-466e-aafb-7a597487f2c0\") " Sep 30 14:24:40 crc kubenswrapper[4840]: I0930 14:24:40.430244 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c378bd2-1562-466e-aafb-7a597487f2c0-kube-api-access-c8gn8" (OuterVolumeSpecName: "kube-api-access-c8gn8") pod "3c378bd2-1562-466e-aafb-7a597487f2c0" (UID: "3c378bd2-1562-466e-aafb-7a597487f2c0"). InnerVolumeSpecName "kube-api-access-c8gn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:24:40 crc kubenswrapper[4840]: I0930 14:24:40.450793 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c378bd2-1562-466e-aafb-7a597487f2c0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3c378bd2-1562-466e-aafb-7a597487f2c0" (UID: "3c378bd2-1562-466e-aafb-7a597487f2c0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:24:40 crc kubenswrapper[4840]: I0930 14:24:40.456869 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c378bd2-1562-466e-aafb-7a597487f2c0-inventory" (OuterVolumeSpecName: "inventory") pod "3c378bd2-1562-466e-aafb-7a597487f2c0" (UID: "3c378bd2-1562-466e-aafb-7a597487f2c0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:24:40 crc kubenswrapper[4840]: I0930 14:24:40.528033 4840 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3c378bd2-1562-466e-aafb-7a597487f2c0-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:24:40 crc kubenswrapper[4840]: I0930 14:24:40.528179 4840 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3c378bd2-1562-466e-aafb-7a597487f2c0-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 14:24:40 crc kubenswrapper[4840]: I0930 14:24:40.528196 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8gn8\" (UniqueName: \"kubernetes.io/projected/3c378bd2-1562-466e-aafb-7a597487f2c0-kube-api-access-c8gn8\") on node \"crc\" DevicePath \"\"" Sep 30 14:24:40 crc kubenswrapper[4840]: I0930 14:24:40.858840 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hw58j" event={"ID":"3c378bd2-1562-466e-aafb-7a597487f2c0","Type":"ContainerDied","Data":"8785f364cf1d9e7efea290d9450dc3caf61c1322604d3e39247290edcae2d3a8"} Sep 30 14:24:40 crc kubenswrapper[4840]: I0930 14:24:40.859131 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8785f364cf1d9e7efea290d9450dc3caf61c1322604d3e39247290edcae2d3a8" Sep 30 14:24:40 crc kubenswrapper[4840]: I0930 14:24:40.858919 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hw58j" Sep 30 14:24:40 crc kubenswrapper[4840]: I0930 14:24:40.935469 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9p2lj"] Sep 30 14:24:40 crc kubenswrapper[4840]: E0930 14:24:40.935871 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee66f4bd-fee1-482e-9eb2-f09f4203084e" containerName="registry-server" Sep 30 14:24:40 crc kubenswrapper[4840]: I0930 14:24:40.935892 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee66f4bd-fee1-482e-9eb2-f09f4203084e" containerName="registry-server" Sep 30 14:24:40 crc kubenswrapper[4840]: E0930 14:24:40.935908 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c378bd2-1562-466e-aafb-7a597487f2c0" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Sep 30 14:24:40 crc kubenswrapper[4840]: I0930 14:24:40.935915 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c378bd2-1562-466e-aafb-7a597487f2c0" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Sep 30 14:24:40 crc kubenswrapper[4840]: E0930 14:24:40.935932 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee66f4bd-fee1-482e-9eb2-f09f4203084e" containerName="extract-utilities" Sep 30 14:24:40 crc kubenswrapper[4840]: I0930 14:24:40.935939 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee66f4bd-fee1-482e-9eb2-f09f4203084e" containerName="extract-utilities" Sep 30 14:24:40 crc kubenswrapper[4840]: E0930 14:24:40.935955 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee66f4bd-fee1-482e-9eb2-f09f4203084e" containerName="extract-content" Sep 30 14:24:40 crc kubenswrapper[4840]: I0930 14:24:40.935961 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee66f4bd-fee1-482e-9eb2-f09f4203084e" containerName="extract-content" Sep 30 14:24:40 crc kubenswrapper[4840]: I0930 14:24:40.936156 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c378bd2-1562-466e-aafb-7a597487f2c0" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Sep 30 14:24:40 crc kubenswrapper[4840]: I0930 14:24:40.936177 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee66f4bd-fee1-482e-9eb2-f09f4203084e" containerName="registry-server" Sep 30 14:24:40 crc kubenswrapper[4840]: I0930 14:24:40.936882 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9p2lj" Sep 30 14:24:40 crc kubenswrapper[4840]: I0930 14:24:40.940256 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 14:24:40 crc kubenswrapper[4840]: I0930 14:24:40.940498 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9zm9z" Sep 30 14:24:40 crc kubenswrapper[4840]: I0930 14:24:40.940702 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 14:24:40 crc kubenswrapper[4840]: I0930 14:24:40.940885 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 14:24:40 crc kubenswrapper[4840]: I0930 14:24:40.959928 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9p2lj"] Sep 30 14:24:41 crc kubenswrapper[4840]: I0930 14:24:41.035489 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxzjw\" (UniqueName: \"kubernetes.io/projected/cbc06ec4-8aae-4192-a201-d2b0d13b29b1-kube-api-access-xxzjw\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9p2lj\" (UID: \"cbc06ec4-8aae-4192-a201-d2b0d13b29b1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9p2lj" Sep 30 14:24:41 crc kubenswrapper[4840]: I0930 14:24:41.036032 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cbc06ec4-8aae-4192-a201-d2b0d13b29b1-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9p2lj\" (UID: \"cbc06ec4-8aae-4192-a201-d2b0d13b29b1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9p2lj" Sep 30 14:24:41 crc kubenswrapper[4840]: I0930 14:24:41.036196 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cbc06ec4-8aae-4192-a201-d2b0d13b29b1-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9p2lj\" (UID: \"cbc06ec4-8aae-4192-a201-d2b0d13b29b1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9p2lj" Sep 30 14:24:41 crc kubenswrapper[4840]: I0930 14:24:41.138654 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cbc06ec4-8aae-4192-a201-d2b0d13b29b1-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9p2lj\" (UID: \"cbc06ec4-8aae-4192-a201-d2b0d13b29b1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9p2lj" Sep 30 14:24:41 crc kubenswrapper[4840]: I0930 14:24:41.138719 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cbc06ec4-8aae-4192-a201-d2b0d13b29b1-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9p2lj\" (UID: \"cbc06ec4-8aae-4192-a201-d2b0d13b29b1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9p2lj" Sep 30 14:24:41 crc kubenswrapper[4840]: I0930 14:24:41.138840 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxzjw\" (UniqueName: \"kubernetes.io/projected/cbc06ec4-8aae-4192-a201-d2b0d13b29b1-kube-api-access-xxzjw\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9p2lj\" (UID: \"cbc06ec4-8aae-4192-a201-d2b0d13b29b1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9p2lj" Sep 30 14:24:41 crc kubenswrapper[4840]: I0930 14:24:41.144601 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cbc06ec4-8aae-4192-a201-d2b0d13b29b1-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9p2lj\" (UID: \"cbc06ec4-8aae-4192-a201-d2b0d13b29b1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9p2lj" Sep 30 14:24:41 crc kubenswrapper[4840]: I0930 14:24:41.145123 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cbc06ec4-8aae-4192-a201-d2b0d13b29b1-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9p2lj\" (UID: \"cbc06ec4-8aae-4192-a201-d2b0d13b29b1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9p2lj" Sep 30 14:24:41 crc kubenswrapper[4840]: I0930 14:24:41.154914 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxzjw\" (UniqueName: \"kubernetes.io/projected/cbc06ec4-8aae-4192-a201-d2b0d13b29b1-kube-api-access-xxzjw\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9p2lj\" (UID: \"cbc06ec4-8aae-4192-a201-d2b0d13b29b1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9p2lj" Sep 30 14:24:41 crc kubenswrapper[4840]: I0930 14:24:41.269402 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9p2lj" Sep 30 14:24:41 crc kubenswrapper[4840]: I0930 14:24:41.797385 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9p2lj"] Sep 30 14:24:41 crc kubenswrapper[4840]: I0930 14:24:41.870511 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9p2lj" event={"ID":"cbc06ec4-8aae-4192-a201-d2b0d13b29b1","Type":"ContainerStarted","Data":"3591011c9e628603dadea6bebb3066c67fc15d241b48af341ed5855f6e636486"} Sep 30 14:24:42 crc kubenswrapper[4840]: I0930 14:24:42.889349 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9p2lj" event={"ID":"cbc06ec4-8aae-4192-a201-d2b0d13b29b1","Type":"ContainerStarted","Data":"3b13fa90ea77832cbf4420de095fa33fda73c050d0447acc090670f84c6aba7b"} Sep 30 14:24:42 crc kubenswrapper[4840]: I0930 14:24:42.913812 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9p2lj" podStartSLOduration=2.412745089 podStartE2EDuration="2.913790392s" podCreationTimestamp="2025-09-30 14:24:40 +0000 UTC" firstStartedPulling="2025-09-30 14:24:41.804016231 +0000 UTC m=+1710.433102654" lastFinishedPulling="2025-09-30 14:24:42.305061534 +0000 UTC m=+1710.934147957" observedRunningTime="2025-09-30 14:24:42.907363868 +0000 UTC m=+1711.536450311" watchObservedRunningTime="2025-09-30 14:24:42.913790392 +0000 UTC m=+1711.542876825" Sep 30 14:24:48 crc kubenswrapper[4840]: I0930 14:24:48.117110 4840 scope.go:117] "RemoveContainer" containerID="74b12cc9318ac99161f1b2541861d89b6100ba9a986e5863f53ae26c2a8953e7" Sep 30 14:24:48 crc kubenswrapper[4840]: E0930 14:24:48.118572 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:24:56 crc kubenswrapper[4840]: I0930 14:24:56.046704 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-nlp8v"] Sep 30 14:24:56 crc kubenswrapper[4840]: I0930 14:24:56.054481 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-nlp8v"] Sep 30 14:24:56 crc kubenswrapper[4840]: I0930 14:24:56.127209 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7e57e20-fd49-4406-9814-1486cc9efdf0" path="/var/lib/kubelet/pods/d7e57e20-fd49-4406-9814-1486cc9efdf0/volumes" Sep 30 14:25:01 crc kubenswrapper[4840]: I0930 14:25:01.116635 4840 scope.go:117] "RemoveContainer" containerID="74b12cc9318ac99161f1b2541861d89b6100ba9a986e5863f53ae26c2a8953e7" Sep 30 14:25:01 crc kubenswrapper[4840]: E0930 14:25:01.117625 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:25:12 crc kubenswrapper[4840]: I0930 14:25:12.125893 4840 scope.go:117] "RemoveContainer" containerID="74b12cc9318ac99161f1b2541861d89b6100ba9a986e5863f53ae26c2a8953e7" Sep 30 14:25:12 crc kubenswrapper[4840]: E0930 14:25:12.127942 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:25:20 crc kubenswrapper[4840]: I0930 14:25:20.041465 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-v9q8t"] Sep 30 14:25:20 crc kubenswrapper[4840]: I0930 14:25:20.048282 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-v9q8t"] Sep 30 14:25:20 crc kubenswrapper[4840]: I0930 14:25:20.135705 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="722796b2-e357-4fec-8437-46f163f510d6" path="/var/lib/kubelet/pods/722796b2-e357-4fec-8437-46f163f510d6/volumes" Sep 30 14:25:25 crc kubenswrapper[4840]: I0930 14:25:25.116533 4840 scope.go:117] "RemoveContainer" containerID="74b12cc9318ac99161f1b2541861d89b6100ba9a986e5863f53ae26c2a8953e7" Sep 30 14:25:25 crc kubenswrapper[4840]: E0930 14:25:25.117474 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:25:34 crc kubenswrapper[4840]: I0930 14:25:34.928096 4840 scope.go:117] "RemoveContainer" containerID="ea758c4f9027095d42744a11ef0b508192d6c55fb24d66b7de02b3df32145dcf" Sep 30 14:25:34 crc kubenswrapper[4840]: I0930 14:25:34.961711 4840 scope.go:117] "RemoveContainer" containerID="8bdf30fd0b762dd4d4417dd4e2b2f0e0e131385af5e4454443d06237bb91c9be" Sep 30 14:25:35 crc kubenswrapper[4840]: I0930 14:25:35.305163 4840 generic.go:334] "Generic (PLEG): container finished" podID="cbc06ec4-8aae-4192-a201-d2b0d13b29b1" containerID="3b13fa90ea77832cbf4420de095fa33fda73c050d0447acc090670f84c6aba7b" exitCode=2 Sep 30 14:25:35 crc kubenswrapper[4840]: I0930 14:25:35.305216 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9p2lj" event={"ID":"cbc06ec4-8aae-4192-a201-d2b0d13b29b1","Type":"ContainerDied","Data":"3b13fa90ea77832cbf4420de095fa33fda73c050d0447acc090670f84c6aba7b"} Sep 30 14:25:36 crc kubenswrapper[4840]: I0930 14:25:36.721523 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9p2lj" Sep 30 14:25:36 crc kubenswrapper[4840]: I0930 14:25:36.795008 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cbc06ec4-8aae-4192-a201-d2b0d13b29b1-ssh-key\") pod \"cbc06ec4-8aae-4192-a201-d2b0d13b29b1\" (UID: \"cbc06ec4-8aae-4192-a201-d2b0d13b29b1\") " Sep 30 14:25:36 crc kubenswrapper[4840]: I0930 14:25:36.795112 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxzjw\" (UniqueName: \"kubernetes.io/projected/cbc06ec4-8aae-4192-a201-d2b0d13b29b1-kube-api-access-xxzjw\") pod \"cbc06ec4-8aae-4192-a201-d2b0d13b29b1\" (UID: \"cbc06ec4-8aae-4192-a201-d2b0d13b29b1\") " Sep 30 14:25:36 crc kubenswrapper[4840]: I0930 14:25:36.795154 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cbc06ec4-8aae-4192-a201-d2b0d13b29b1-inventory\") pod \"cbc06ec4-8aae-4192-a201-d2b0d13b29b1\" (UID: \"cbc06ec4-8aae-4192-a201-d2b0d13b29b1\") " Sep 30 14:25:36 crc kubenswrapper[4840]: I0930 14:25:36.800730 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbc06ec4-8aae-4192-a201-d2b0d13b29b1-kube-api-access-xxzjw" (OuterVolumeSpecName: "kube-api-access-xxzjw") pod "cbc06ec4-8aae-4192-a201-d2b0d13b29b1" (UID: "cbc06ec4-8aae-4192-a201-d2b0d13b29b1"). InnerVolumeSpecName "kube-api-access-xxzjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:25:36 crc kubenswrapper[4840]: I0930 14:25:36.825122 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbc06ec4-8aae-4192-a201-d2b0d13b29b1-inventory" (OuterVolumeSpecName: "inventory") pod "cbc06ec4-8aae-4192-a201-d2b0d13b29b1" (UID: "cbc06ec4-8aae-4192-a201-d2b0d13b29b1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:25:36 crc kubenswrapper[4840]: I0930 14:25:36.826615 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbc06ec4-8aae-4192-a201-d2b0d13b29b1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "cbc06ec4-8aae-4192-a201-d2b0d13b29b1" (UID: "cbc06ec4-8aae-4192-a201-d2b0d13b29b1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:25:36 crc kubenswrapper[4840]: I0930 14:25:36.897839 4840 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cbc06ec4-8aae-4192-a201-d2b0d13b29b1-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:25:36 crc kubenswrapper[4840]: I0930 14:25:36.897888 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxzjw\" (UniqueName: \"kubernetes.io/projected/cbc06ec4-8aae-4192-a201-d2b0d13b29b1-kube-api-access-xxzjw\") on node \"crc\" DevicePath \"\"" Sep 30 14:25:36 crc kubenswrapper[4840]: I0930 14:25:36.897901 4840 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cbc06ec4-8aae-4192-a201-d2b0d13b29b1-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 14:25:37 crc kubenswrapper[4840]: I0930 14:25:37.328297 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9p2lj" event={"ID":"cbc06ec4-8aae-4192-a201-d2b0d13b29b1","Type":"ContainerDied","Data":"3591011c9e628603dadea6bebb3066c67fc15d241b48af341ed5855f6e636486"} Sep 30 14:25:37 crc kubenswrapper[4840]: I0930 14:25:37.328346 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3591011c9e628603dadea6bebb3066c67fc15d241b48af341ed5855f6e636486" Sep 30 14:25:37 crc kubenswrapper[4840]: I0930 14:25:37.328365 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9p2lj" Sep 30 14:25:40 crc kubenswrapper[4840]: I0930 14:25:40.116243 4840 scope.go:117] "RemoveContainer" containerID="74b12cc9318ac99161f1b2541861d89b6100ba9a986e5863f53ae26c2a8953e7" Sep 30 14:25:40 crc kubenswrapper[4840]: E0930 14:25:40.117085 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:25:44 crc kubenswrapper[4840]: I0930 14:25:44.028241 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bmz4g"] Sep 30 14:25:44 crc kubenswrapper[4840]: E0930 14:25:44.029207 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbc06ec4-8aae-4192-a201-d2b0d13b29b1" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Sep 30 14:25:44 crc kubenswrapper[4840]: I0930 14:25:44.029230 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbc06ec4-8aae-4192-a201-d2b0d13b29b1" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Sep 30 14:25:44 crc kubenswrapper[4840]: I0930 14:25:44.029442 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbc06ec4-8aae-4192-a201-d2b0d13b29b1" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Sep 30 14:25:44 crc kubenswrapper[4840]: I0930 14:25:44.030320 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bmz4g" Sep 30 14:25:44 crc kubenswrapper[4840]: I0930 14:25:44.033666 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 14:25:44 crc kubenswrapper[4840]: I0930 14:25:44.035137 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9zm9z" Sep 30 14:25:44 crc kubenswrapper[4840]: I0930 14:25:44.035482 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 14:25:44 crc kubenswrapper[4840]: I0930 14:25:44.035675 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 14:25:44 crc kubenswrapper[4840]: I0930 14:25:44.042825 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bmz4g"] Sep 30 14:25:44 crc kubenswrapper[4840]: I0930 14:25:44.142270 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6nnw\" (UniqueName: \"kubernetes.io/projected/797b0ada-6af8-4509-8a81-6b97c263fdc3-kube-api-access-k6nnw\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bmz4g\" (UID: \"797b0ada-6af8-4509-8a81-6b97c263fdc3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bmz4g" Sep 30 14:25:44 crc kubenswrapper[4840]: I0930 14:25:44.142388 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/797b0ada-6af8-4509-8a81-6b97c263fdc3-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bmz4g\" (UID: \"797b0ada-6af8-4509-8a81-6b97c263fdc3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bmz4g" Sep 30 14:25:44 crc kubenswrapper[4840]: I0930 14:25:44.142449 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/797b0ada-6af8-4509-8a81-6b97c263fdc3-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bmz4g\" (UID: \"797b0ada-6af8-4509-8a81-6b97c263fdc3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bmz4g" Sep 30 14:25:44 crc kubenswrapper[4840]: I0930 14:25:44.244218 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6nnw\" (UniqueName: \"kubernetes.io/projected/797b0ada-6af8-4509-8a81-6b97c263fdc3-kube-api-access-k6nnw\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bmz4g\" (UID: \"797b0ada-6af8-4509-8a81-6b97c263fdc3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bmz4g" Sep 30 14:25:44 crc kubenswrapper[4840]: I0930 14:25:44.244375 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/797b0ada-6af8-4509-8a81-6b97c263fdc3-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bmz4g\" (UID: \"797b0ada-6af8-4509-8a81-6b97c263fdc3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bmz4g" Sep 30 14:25:44 crc kubenswrapper[4840]: I0930 14:25:44.244425 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/797b0ada-6af8-4509-8a81-6b97c263fdc3-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bmz4g\" (UID: \"797b0ada-6af8-4509-8a81-6b97c263fdc3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bmz4g" Sep 30 14:25:44 crc kubenswrapper[4840]: I0930 14:25:44.251170 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/797b0ada-6af8-4509-8a81-6b97c263fdc3-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bmz4g\" (UID: \"797b0ada-6af8-4509-8a81-6b97c263fdc3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bmz4g" Sep 30 14:25:44 crc kubenswrapper[4840]: I0930 14:25:44.251226 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/797b0ada-6af8-4509-8a81-6b97c263fdc3-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bmz4g\" (UID: \"797b0ada-6af8-4509-8a81-6b97c263fdc3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bmz4g" Sep 30 14:25:44 crc kubenswrapper[4840]: I0930 14:25:44.264054 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6nnw\" (UniqueName: \"kubernetes.io/projected/797b0ada-6af8-4509-8a81-6b97c263fdc3-kube-api-access-k6nnw\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bmz4g\" (UID: \"797b0ada-6af8-4509-8a81-6b97c263fdc3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bmz4g" Sep 30 14:25:44 crc kubenswrapper[4840]: I0930 14:25:44.361189 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bmz4g" Sep 30 14:25:44 crc kubenswrapper[4840]: I0930 14:25:44.868110 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bmz4g"] Sep 30 14:25:45 crc kubenswrapper[4840]: I0930 14:25:45.408542 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bmz4g" event={"ID":"797b0ada-6af8-4509-8a81-6b97c263fdc3","Type":"ContainerStarted","Data":"37f0de64d775304e8c5402856bf15010f65b627c44dc8a01391c54cdf3b1fa10"} Sep 30 14:25:46 crc kubenswrapper[4840]: I0930 14:25:46.418963 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bmz4g" event={"ID":"797b0ada-6af8-4509-8a81-6b97c263fdc3","Type":"ContainerStarted","Data":"4810ecb752d2dd7beb411ba4c7720260dd9ffb81c7693a772c5d9f8721478e11"} Sep 30 14:25:46 crc kubenswrapper[4840]: I0930 14:25:46.440220 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bmz4g" podStartSLOduration=2.016179378 podStartE2EDuration="2.440204946s" podCreationTimestamp="2025-09-30 14:25:44 +0000 UTC" firstStartedPulling="2025-09-30 14:25:44.879232497 +0000 UTC m=+1773.508318920" lastFinishedPulling="2025-09-30 14:25:45.303258065 +0000 UTC m=+1773.932344488" observedRunningTime="2025-09-30 14:25:46.439216641 +0000 UTC m=+1775.068303064" watchObservedRunningTime="2025-09-30 14:25:46.440204946 +0000 UTC m=+1775.069291369" Sep 30 14:25:54 crc kubenswrapper[4840]: I0930 14:25:54.116496 4840 scope.go:117] "RemoveContainer" containerID="74b12cc9318ac99161f1b2541861d89b6100ba9a986e5863f53ae26c2a8953e7" Sep 30 14:25:54 crc kubenswrapper[4840]: E0930 14:25:54.117375 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:26:05 crc kubenswrapper[4840]: I0930 14:26:05.117126 4840 scope.go:117] "RemoveContainer" containerID="74b12cc9318ac99161f1b2541861d89b6100ba9a986e5863f53ae26c2a8953e7" Sep 30 14:26:05 crc kubenswrapper[4840]: E0930 14:26:05.118304 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:26:10 crc kubenswrapper[4840]: I0930 14:26:10.035997 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-2ph7k"] Sep 30 14:26:10 crc kubenswrapper[4840]: I0930 14:26:10.042404 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-2ph7k"] Sep 30 14:26:10 crc kubenswrapper[4840]: I0930 14:26:10.134376 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91fc410c-6ef8-4864-8ba2-a604c7bcd8ac" path="/var/lib/kubelet/pods/91fc410c-6ef8-4864-8ba2-a604c7bcd8ac/volumes" Sep 30 14:26:19 crc kubenswrapper[4840]: I0930 14:26:19.116887 4840 scope.go:117] "RemoveContainer" containerID="74b12cc9318ac99161f1b2541861d89b6100ba9a986e5863f53ae26c2a8953e7" Sep 30 14:26:19 crc kubenswrapper[4840]: E0930 14:26:19.117813 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:26:25 crc kubenswrapper[4840]: I0930 14:26:25.030980 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-f9pkg"] Sep 30 14:26:25 crc kubenswrapper[4840]: I0930 14:26:25.039120 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-f9pkg"] Sep 30 14:26:26 crc kubenswrapper[4840]: I0930 14:26:26.130478 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="521c6360-ed9d-4b6d-a74b-188e3a27be00" path="/var/lib/kubelet/pods/521c6360-ed9d-4b6d-a74b-188e3a27be00/volumes" Sep 30 14:26:28 crc kubenswrapper[4840]: I0930 14:26:28.803457 4840 generic.go:334] "Generic (PLEG): container finished" podID="797b0ada-6af8-4509-8a81-6b97c263fdc3" containerID="4810ecb752d2dd7beb411ba4c7720260dd9ffb81c7693a772c5d9f8721478e11" exitCode=0 Sep 30 14:26:28 crc kubenswrapper[4840]: I0930 14:26:28.803546 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bmz4g" event={"ID":"797b0ada-6af8-4509-8a81-6b97c263fdc3","Type":"ContainerDied","Data":"4810ecb752d2dd7beb411ba4c7720260dd9ffb81c7693a772c5d9f8721478e11"} Sep 30 14:26:30 crc kubenswrapper[4840]: I0930 14:26:30.237872 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bmz4g" Sep 30 14:26:30 crc kubenswrapper[4840]: I0930 14:26:30.282128 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/797b0ada-6af8-4509-8a81-6b97c263fdc3-inventory\") pod \"797b0ada-6af8-4509-8a81-6b97c263fdc3\" (UID: \"797b0ada-6af8-4509-8a81-6b97c263fdc3\") " Sep 30 14:26:30 crc kubenswrapper[4840]: I0930 14:26:30.282218 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6nnw\" (UniqueName: \"kubernetes.io/projected/797b0ada-6af8-4509-8a81-6b97c263fdc3-kube-api-access-k6nnw\") pod \"797b0ada-6af8-4509-8a81-6b97c263fdc3\" (UID: \"797b0ada-6af8-4509-8a81-6b97c263fdc3\") " Sep 30 14:26:30 crc kubenswrapper[4840]: I0930 14:26:30.282285 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/797b0ada-6af8-4509-8a81-6b97c263fdc3-ssh-key\") pod \"797b0ada-6af8-4509-8a81-6b97c263fdc3\" (UID: \"797b0ada-6af8-4509-8a81-6b97c263fdc3\") " Sep 30 14:26:30 crc kubenswrapper[4840]: I0930 14:26:30.295756 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/797b0ada-6af8-4509-8a81-6b97c263fdc3-kube-api-access-k6nnw" (OuterVolumeSpecName: "kube-api-access-k6nnw") pod "797b0ada-6af8-4509-8a81-6b97c263fdc3" (UID: "797b0ada-6af8-4509-8a81-6b97c263fdc3"). InnerVolumeSpecName "kube-api-access-k6nnw". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:26:30 crc kubenswrapper[4840]: I0930 14:26:30.379760 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/797b0ada-6af8-4509-8a81-6b97c263fdc3-inventory" (OuterVolumeSpecName: "inventory") pod "797b0ada-6af8-4509-8a81-6b97c263fdc3" (UID: "797b0ada-6af8-4509-8a81-6b97c263fdc3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:26:30 crc kubenswrapper[4840]: I0930 14:26:30.387836 4840 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/797b0ada-6af8-4509-8a81-6b97c263fdc3-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 14:26:30 crc kubenswrapper[4840]: I0930 14:26:30.387863 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6nnw\" (UniqueName: \"kubernetes.io/projected/797b0ada-6af8-4509-8a81-6b97c263fdc3-kube-api-access-k6nnw\") on node \"crc\" DevicePath \"\"" Sep 30 14:26:30 crc kubenswrapper[4840]: I0930 14:26:30.422793 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/797b0ada-6af8-4509-8a81-6b97c263fdc3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "797b0ada-6af8-4509-8a81-6b97c263fdc3" (UID: "797b0ada-6af8-4509-8a81-6b97c263fdc3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:26:30 crc kubenswrapper[4840]: I0930 14:26:30.489294 4840 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/797b0ada-6af8-4509-8a81-6b97c263fdc3-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:26:30 crc kubenswrapper[4840]: I0930 14:26:30.822968 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bmz4g" event={"ID":"797b0ada-6af8-4509-8a81-6b97c263fdc3","Type":"ContainerDied","Data":"37f0de64d775304e8c5402856bf15010f65b627c44dc8a01391c54cdf3b1fa10"} Sep 30 14:26:30 crc kubenswrapper[4840]: I0930 14:26:30.823012 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37f0de64d775304e8c5402856bf15010f65b627c44dc8a01391c54cdf3b1fa10" Sep 30 14:26:30 crc kubenswrapper[4840]: I0930 14:26:30.823033 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bmz4g" Sep 30 14:26:30 crc kubenswrapper[4840]: I0930 14:26:30.915350 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-dbvwh"] Sep 30 14:26:30 crc kubenswrapper[4840]: E0930 14:26:30.915791 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="797b0ada-6af8-4509-8a81-6b97c263fdc3" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Sep 30 14:26:30 crc kubenswrapper[4840]: I0930 14:26:30.915816 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="797b0ada-6af8-4509-8a81-6b97c263fdc3" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Sep 30 14:26:30 crc kubenswrapper[4840]: I0930 14:26:30.916070 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="797b0ada-6af8-4509-8a81-6b97c263fdc3" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Sep 30 14:26:30 crc kubenswrapper[4840]: I0930 14:26:30.916836 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-dbvwh" Sep 30 14:26:30 crc kubenswrapper[4840]: I0930 14:26:30.920732 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 14:26:30 crc kubenswrapper[4840]: I0930 14:26:30.920981 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 14:26:30 crc kubenswrapper[4840]: I0930 14:26:30.921709 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 14:26:30 crc kubenswrapper[4840]: I0930 14:26:30.925710 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9zm9z" Sep 30 14:26:30 crc kubenswrapper[4840]: I0930 14:26:30.933860 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-dbvwh"] Sep 30 14:26:30 crc kubenswrapper[4840]: I0930 14:26:30.998765 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qg66j\" (UniqueName: \"kubernetes.io/projected/37a41ba6-72d7-40d8-a305-153b68f100a8-kube-api-access-qg66j\") pod \"ssh-known-hosts-edpm-deployment-dbvwh\" (UID: \"37a41ba6-72d7-40d8-a305-153b68f100a8\") " pod="openstack/ssh-known-hosts-edpm-deployment-dbvwh" Sep 30 14:26:30 crc kubenswrapper[4840]: I0930 14:26:30.998885 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/37a41ba6-72d7-40d8-a305-153b68f100a8-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-dbvwh\" (UID: \"37a41ba6-72d7-40d8-a305-153b68f100a8\") " pod="openstack/ssh-known-hosts-edpm-deployment-dbvwh" Sep 30 14:26:30 crc kubenswrapper[4840]: I0930 14:26:30.998959 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/37a41ba6-72d7-40d8-a305-153b68f100a8-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-dbvwh\" (UID: \"37a41ba6-72d7-40d8-a305-153b68f100a8\") " pod="openstack/ssh-known-hosts-edpm-deployment-dbvwh" Sep 30 14:26:31 crc kubenswrapper[4840]: I0930 14:26:31.100710 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/37a41ba6-72d7-40d8-a305-153b68f100a8-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-dbvwh\" (UID: \"37a41ba6-72d7-40d8-a305-153b68f100a8\") " pod="openstack/ssh-known-hosts-edpm-deployment-dbvwh" Sep 30 14:26:31 crc kubenswrapper[4840]: I0930 14:26:31.100832 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qg66j\" (UniqueName: \"kubernetes.io/projected/37a41ba6-72d7-40d8-a305-153b68f100a8-kube-api-access-qg66j\") pod \"ssh-known-hosts-edpm-deployment-dbvwh\" (UID: \"37a41ba6-72d7-40d8-a305-153b68f100a8\") " pod="openstack/ssh-known-hosts-edpm-deployment-dbvwh" Sep 30 14:26:31 crc kubenswrapper[4840]: I0930 14:26:31.100894 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/37a41ba6-72d7-40d8-a305-153b68f100a8-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-dbvwh\" (UID: \"37a41ba6-72d7-40d8-a305-153b68f100a8\") " pod="openstack/ssh-known-hosts-edpm-deployment-dbvwh" Sep 30 14:26:31 crc kubenswrapper[4840]: I0930 14:26:31.104336 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/37a41ba6-72d7-40d8-a305-153b68f100a8-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-dbvwh\" (UID: \"37a41ba6-72d7-40d8-a305-153b68f100a8\") " pod="openstack/ssh-known-hosts-edpm-deployment-dbvwh" Sep 30 14:26:31 crc kubenswrapper[4840]: I0930 14:26:31.104436 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/37a41ba6-72d7-40d8-a305-153b68f100a8-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-dbvwh\" (UID: \"37a41ba6-72d7-40d8-a305-153b68f100a8\") " pod="openstack/ssh-known-hosts-edpm-deployment-dbvwh" Sep 30 14:26:31 crc kubenswrapper[4840]: I0930 14:26:31.116638 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qg66j\" (UniqueName: \"kubernetes.io/projected/37a41ba6-72d7-40d8-a305-153b68f100a8-kube-api-access-qg66j\") pod \"ssh-known-hosts-edpm-deployment-dbvwh\" (UID: \"37a41ba6-72d7-40d8-a305-153b68f100a8\") " pod="openstack/ssh-known-hosts-edpm-deployment-dbvwh" Sep 30 14:26:31 crc kubenswrapper[4840]: I0930 14:26:31.236954 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-dbvwh" Sep 30 14:26:31 crc kubenswrapper[4840]: I0930 14:26:31.743977 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-dbvwh"] Sep 30 14:26:31 crc kubenswrapper[4840]: I0930 14:26:31.831862 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-dbvwh" event={"ID":"37a41ba6-72d7-40d8-a305-153b68f100a8","Type":"ContainerStarted","Data":"712f24e0fe5bf38555295938df4fb1b264555852a7024488fe66cf4b0e56e15a"} Sep 30 14:26:33 crc kubenswrapper[4840]: I0930 14:26:33.853785 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-dbvwh" event={"ID":"37a41ba6-72d7-40d8-a305-153b68f100a8","Type":"ContainerStarted","Data":"35307466cec5713cb0dc2879d458c3bac6c8d930ce2050bc7ccae5f4274ce707"} Sep 30 14:26:33 crc kubenswrapper[4840]: I0930 14:26:33.880041 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-dbvwh" podStartSLOduration=3.047952404 podStartE2EDuration="3.880019819s" podCreationTimestamp="2025-09-30 14:26:30 +0000 UTC" firstStartedPulling="2025-09-30 14:26:31.747226344 +0000 UTC m=+1820.376312767" lastFinishedPulling="2025-09-30 14:26:32.579293749 +0000 UTC m=+1821.208380182" observedRunningTime="2025-09-30 14:26:33.873187674 +0000 UTC m=+1822.502274107" watchObservedRunningTime="2025-09-30 14:26:33.880019819 +0000 UTC m=+1822.509106262" Sep 30 14:26:34 crc kubenswrapper[4840]: I0930 14:26:34.116399 4840 scope.go:117] "RemoveContainer" containerID="74b12cc9318ac99161f1b2541861d89b6100ba9a986e5863f53ae26c2a8953e7" Sep 30 14:26:34 crc kubenswrapper[4840]: E0930 14:26:34.116722 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:26:35 crc kubenswrapper[4840]: I0930 14:26:35.051073 4840 scope.go:117] "RemoveContainer" containerID="beb1299b9489e95f9a0952cc2035d97810aeeae6821fb33d18934840d4a0b784" Sep 30 14:26:35 crc kubenswrapper[4840]: I0930 14:26:35.094186 4840 scope.go:117] "RemoveContainer" containerID="d2a2e49e9d9bfd23804ae23d4fe6675ebfe92252517fe6c60b4fc629f086bff7" Sep 30 14:26:39 crc kubenswrapper[4840]: I0930 14:26:39.905359 4840 generic.go:334] "Generic (PLEG): container finished" podID="37a41ba6-72d7-40d8-a305-153b68f100a8" containerID="35307466cec5713cb0dc2879d458c3bac6c8d930ce2050bc7ccae5f4274ce707" exitCode=0 Sep 30 14:26:39 crc kubenswrapper[4840]: I0930 14:26:39.905434 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-dbvwh" event={"ID":"37a41ba6-72d7-40d8-a305-153b68f100a8","Type":"ContainerDied","Data":"35307466cec5713cb0dc2879d458c3bac6c8d930ce2050bc7ccae5f4274ce707"} Sep 30 14:26:41 crc kubenswrapper[4840]: I0930 14:26:41.309246 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-dbvwh" Sep 30 14:26:41 crc kubenswrapper[4840]: I0930 14:26:41.390648 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg66j\" (UniqueName: \"kubernetes.io/projected/37a41ba6-72d7-40d8-a305-153b68f100a8-kube-api-access-qg66j\") pod \"37a41ba6-72d7-40d8-a305-153b68f100a8\" (UID: \"37a41ba6-72d7-40d8-a305-153b68f100a8\") " Sep 30 14:26:41 crc kubenswrapper[4840]: I0930 14:26:41.391075 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/37a41ba6-72d7-40d8-a305-153b68f100a8-inventory-0\") pod \"37a41ba6-72d7-40d8-a305-153b68f100a8\" (UID: \"37a41ba6-72d7-40d8-a305-153b68f100a8\") " Sep 30 14:26:41 crc kubenswrapper[4840]: I0930 14:26:41.391248 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/37a41ba6-72d7-40d8-a305-153b68f100a8-ssh-key-openstack-edpm-ipam\") pod \"37a41ba6-72d7-40d8-a305-153b68f100a8\" (UID: \"37a41ba6-72d7-40d8-a305-153b68f100a8\") " Sep 30 14:26:41 crc kubenswrapper[4840]: I0930 14:26:41.396182 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37a41ba6-72d7-40d8-a305-153b68f100a8-kube-api-access-qg66j" (OuterVolumeSpecName: "kube-api-access-qg66j") pod "37a41ba6-72d7-40d8-a305-153b68f100a8" (UID: "37a41ba6-72d7-40d8-a305-153b68f100a8"). InnerVolumeSpecName "kube-api-access-qg66j". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:26:41 crc kubenswrapper[4840]: I0930 14:26:41.417988 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37a41ba6-72d7-40d8-a305-153b68f100a8-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "37a41ba6-72d7-40d8-a305-153b68f100a8" (UID: "37a41ba6-72d7-40d8-a305-153b68f100a8"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:26:41 crc kubenswrapper[4840]: I0930 14:26:41.421763 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37a41ba6-72d7-40d8-a305-153b68f100a8-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "37a41ba6-72d7-40d8-a305-153b68f100a8" (UID: "37a41ba6-72d7-40d8-a305-153b68f100a8"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:26:41 crc kubenswrapper[4840]: I0930 14:26:41.493392 4840 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/37a41ba6-72d7-40d8-a305-153b68f100a8-inventory-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:26:41 crc kubenswrapper[4840]: I0930 14:26:41.493420 4840 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/37a41ba6-72d7-40d8-a305-153b68f100a8-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Sep 30 14:26:41 crc kubenswrapper[4840]: I0930 14:26:41.493438 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg66j\" (UniqueName: \"kubernetes.io/projected/37a41ba6-72d7-40d8-a305-153b68f100a8-kube-api-access-qg66j\") on node \"crc\" DevicePath \"\"" Sep 30 14:26:41 crc kubenswrapper[4840]: I0930 14:26:41.922543 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-dbvwh" Sep 30 14:26:41 crc kubenswrapper[4840]: I0930 14:26:41.922509 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-dbvwh" event={"ID":"37a41ba6-72d7-40d8-a305-153b68f100a8","Type":"ContainerDied","Data":"712f24e0fe5bf38555295938df4fb1b264555852a7024488fe66cf4b0e56e15a"} Sep 30 14:26:41 crc kubenswrapper[4840]: I0930 14:26:41.922610 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="712f24e0fe5bf38555295938df4fb1b264555852a7024488fe66cf4b0e56e15a" Sep 30 14:26:41 crc kubenswrapper[4840]: I0930 14:26:41.982307 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-nn4ls"] Sep 30 14:26:41 crc kubenswrapper[4840]: E0930 14:26:41.982803 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37a41ba6-72d7-40d8-a305-153b68f100a8" containerName="ssh-known-hosts-edpm-deployment" Sep 30 14:26:41 crc kubenswrapper[4840]: I0930 14:26:41.982830 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="37a41ba6-72d7-40d8-a305-153b68f100a8" containerName="ssh-known-hosts-edpm-deployment" Sep 30 14:26:41 crc kubenswrapper[4840]: I0930 14:26:41.983055 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="37a41ba6-72d7-40d8-a305-153b68f100a8" containerName="ssh-known-hosts-edpm-deployment" Sep 30 14:26:41 crc kubenswrapper[4840]: I0930 14:26:41.983906 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-nn4ls" Sep 30 14:26:41 crc kubenswrapper[4840]: I0930 14:26:41.988146 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 14:26:41 crc kubenswrapper[4840]: I0930 14:26:41.988432 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9zm9z" Sep 30 14:26:41 crc kubenswrapper[4840]: I0930 14:26:41.988602 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 14:26:41 crc kubenswrapper[4840]: I0930 14:26:41.988754 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 14:26:41 crc kubenswrapper[4840]: I0930 14:26:41.998288 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-nn4ls"] Sep 30 14:26:42 crc kubenswrapper[4840]: I0930 14:26:42.105375 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7b69c13-1a32-4676-9912-e117b0b59848-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-nn4ls\" (UID: \"a7b69c13-1a32-4676-9912-e117b0b59848\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-nn4ls" Sep 30 14:26:42 crc kubenswrapper[4840]: I0930 14:26:42.105463 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a7b69c13-1a32-4676-9912-e117b0b59848-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-nn4ls\" (UID: \"a7b69c13-1a32-4676-9912-e117b0b59848\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-nn4ls" Sep 30 14:26:42 crc kubenswrapper[4840]: I0930 14:26:42.105863 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5cmf\" (UniqueName: \"kubernetes.io/projected/a7b69c13-1a32-4676-9912-e117b0b59848-kube-api-access-n5cmf\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-nn4ls\" (UID: \"a7b69c13-1a32-4676-9912-e117b0b59848\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-nn4ls" Sep 30 14:26:42 crc kubenswrapper[4840]: I0930 14:26:42.207531 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7b69c13-1a32-4676-9912-e117b0b59848-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-nn4ls\" (UID: \"a7b69c13-1a32-4676-9912-e117b0b59848\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-nn4ls" Sep 30 14:26:42 crc kubenswrapper[4840]: I0930 14:26:42.207649 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a7b69c13-1a32-4676-9912-e117b0b59848-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-nn4ls\" (UID: \"a7b69c13-1a32-4676-9912-e117b0b59848\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-nn4ls" Sep 30 14:26:42 crc kubenswrapper[4840]: I0930 14:26:42.207745 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5cmf\" (UniqueName: \"kubernetes.io/projected/a7b69c13-1a32-4676-9912-e117b0b59848-kube-api-access-n5cmf\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-nn4ls\" (UID: \"a7b69c13-1a32-4676-9912-e117b0b59848\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-nn4ls" Sep 30 14:26:42 crc kubenswrapper[4840]: I0930 14:26:42.211791 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a7b69c13-1a32-4676-9912-e117b0b59848-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-nn4ls\" (UID: \"a7b69c13-1a32-4676-9912-e117b0b59848\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-nn4ls" Sep 30 14:26:42 crc kubenswrapper[4840]: I0930 14:26:42.212261 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7b69c13-1a32-4676-9912-e117b0b59848-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-nn4ls\" (UID: \"a7b69c13-1a32-4676-9912-e117b0b59848\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-nn4ls" Sep 30 14:26:42 crc kubenswrapper[4840]: I0930 14:26:42.226309 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5cmf\" (UniqueName: \"kubernetes.io/projected/a7b69c13-1a32-4676-9912-e117b0b59848-kube-api-access-n5cmf\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-nn4ls\" (UID: \"a7b69c13-1a32-4676-9912-e117b0b59848\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-nn4ls" Sep 30 14:26:42 crc kubenswrapper[4840]: I0930 14:26:42.304874 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-nn4ls" Sep 30 14:26:42 crc kubenswrapper[4840]: I0930 14:26:42.803896 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-nn4ls"] Sep 30 14:26:42 crc kubenswrapper[4840]: I0930 14:26:42.930707 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-nn4ls" event={"ID":"a7b69c13-1a32-4676-9912-e117b0b59848","Type":"ContainerStarted","Data":"5bd4322c38e9a09398790bd29848b4601d4bde74cc3b79f18c351097659e9590"} Sep 30 14:26:43 crc kubenswrapper[4840]: I0930 14:26:43.939752 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-nn4ls" event={"ID":"a7b69c13-1a32-4676-9912-e117b0b59848","Type":"ContainerStarted","Data":"e00ee83b31b35292001bcdf0b75158d88e6c82ce4e7d4500d6832858d5fd57ae"} Sep 30 14:26:43 crc kubenswrapper[4840]: I0930 14:26:43.966500 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-nn4ls" podStartSLOduration=2.447731103 podStartE2EDuration="2.966486286s" podCreationTimestamp="2025-09-30 14:26:41 +0000 UTC" firstStartedPulling="2025-09-30 14:26:42.813615956 +0000 UTC m=+1831.442702379" lastFinishedPulling="2025-09-30 14:26:43.332371139 +0000 UTC m=+1831.961457562" observedRunningTime="2025-09-30 14:26:43.955020832 +0000 UTC m=+1832.584107265" watchObservedRunningTime="2025-09-30 14:26:43.966486286 +0000 UTC m=+1832.595572709" Sep 30 14:26:45 crc kubenswrapper[4840]: I0930 14:26:45.117040 4840 scope.go:117] "RemoveContainer" containerID="74b12cc9318ac99161f1b2541861d89b6100ba9a986e5863f53ae26c2a8953e7" Sep 30 14:26:45 crc kubenswrapper[4840]: E0930 14:26:45.117606 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:26:52 crc kubenswrapper[4840]: I0930 14:26:52.013658 4840 generic.go:334] "Generic (PLEG): container finished" podID="a7b69c13-1a32-4676-9912-e117b0b59848" containerID="e00ee83b31b35292001bcdf0b75158d88e6c82ce4e7d4500d6832858d5fd57ae" exitCode=0 Sep 30 14:26:52 crc kubenswrapper[4840]: I0930 14:26:52.013777 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-nn4ls" event={"ID":"a7b69c13-1a32-4676-9912-e117b0b59848","Type":"ContainerDied","Data":"e00ee83b31b35292001bcdf0b75158d88e6c82ce4e7d4500d6832858d5fd57ae"} Sep 30 14:26:53 crc kubenswrapper[4840]: I0930 14:26:53.403056 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-nn4ls" Sep 30 14:26:53 crc kubenswrapper[4840]: I0930 14:26:53.521320 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5cmf\" (UniqueName: \"kubernetes.io/projected/a7b69c13-1a32-4676-9912-e117b0b59848-kube-api-access-n5cmf\") pod \"a7b69c13-1a32-4676-9912-e117b0b59848\" (UID: \"a7b69c13-1a32-4676-9912-e117b0b59848\") " Sep 30 14:26:53 crc kubenswrapper[4840]: I0930 14:26:53.521449 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a7b69c13-1a32-4676-9912-e117b0b59848-ssh-key\") pod \"a7b69c13-1a32-4676-9912-e117b0b59848\" (UID: \"a7b69c13-1a32-4676-9912-e117b0b59848\") " Sep 30 14:26:53 crc kubenswrapper[4840]: I0930 14:26:53.521504 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7b69c13-1a32-4676-9912-e117b0b59848-inventory\") pod \"a7b69c13-1a32-4676-9912-e117b0b59848\" (UID: \"a7b69c13-1a32-4676-9912-e117b0b59848\") " Sep 30 14:26:53 crc kubenswrapper[4840]: I0930 14:26:53.526990 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7b69c13-1a32-4676-9912-e117b0b59848-kube-api-access-n5cmf" (OuterVolumeSpecName: "kube-api-access-n5cmf") pod "a7b69c13-1a32-4676-9912-e117b0b59848" (UID: "a7b69c13-1a32-4676-9912-e117b0b59848"). InnerVolumeSpecName "kube-api-access-n5cmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:26:53 crc kubenswrapper[4840]: I0930 14:26:53.550350 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7b69c13-1a32-4676-9912-e117b0b59848-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a7b69c13-1a32-4676-9912-e117b0b59848" (UID: "a7b69c13-1a32-4676-9912-e117b0b59848"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:26:53 crc kubenswrapper[4840]: I0930 14:26:53.551982 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7b69c13-1a32-4676-9912-e117b0b59848-inventory" (OuterVolumeSpecName: "inventory") pod "a7b69c13-1a32-4676-9912-e117b0b59848" (UID: "a7b69c13-1a32-4676-9912-e117b0b59848"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:26:53 crc kubenswrapper[4840]: I0930 14:26:53.624170 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5cmf\" (UniqueName: \"kubernetes.io/projected/a7b69c13-1a32-4676-9912-e117b0b59848-kube-api-access-n5cmf\") on node \"crc\" DevicePath \"\"" Sep 30 14:26:53 crc kubenswrapper[4840]: I0930 14:26:53.624223 4840 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a7b69c13-1a32-4676-9912-e117b0b59848-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:26:53 crc kubenswrapper[4840]: I0930 14:26:53.624237 4840 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7b69c13-1a32-4676-9912-e117b0b59848-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 14:26:54 crc kubenswrapper[4840]: I0930 14:26:54.032288 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-nn4ls" event={"ID":"a7b69c13-1a32-4676-9912-e117b0b59848","Type":"ContainerDied","Data":"5bd4322c38e9a09398790bd29848b4601d4bde74cc3b79f18c351097659e9590"} Sep 30 14:26:54 crc kubenswrapper[4840]: I0930 14:26:54.032605 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5bd4322c38e9a09398790bd29848b4601d4bde74cc3b79f18c351097659e9590" Sep 30 14:26:54 crc kubenswrapper[4840]: I0930 14:26:54.032387 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-nn4ls" Sep 30 14:26:54 crc kubenswrapper[4840]: I0930 14:26:54.104546 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xkx5r"] Sep 30 14:26:54 crc kubenswrapper[4840]: E0930 14:26:54.104993 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7b69c13-1a32-4676-9912-e117b0b59848" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Sep 30 14:26:54 crc kubenswrapper[4840]: I0930 14:26:54.105012 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7b69c13-1a32-4676-9912-e117b0b59848" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Sep 30 14:26:54 crc kubenswrapper[4840]: I0930 14:26:54.105245 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7b69c13-1a32-4676-9912-e117b0b59848" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Sep 30 14:26:54 crc kubenswrapper[4840]: I0930 14:26:54.106032 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xkx5r" Sep 30 14:26:54 crc kubenswrapper[4840]: I0930 14:26:54.117993 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 14:26:54 crc kubenswrapper[4840]: I0930 14:26:54.118079 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 14:26:54 crc kubenswrapper[4840]: I0930 14:26:54.119589 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 14:26:54 crc kubenswrapper[4840]: I0930 14:26:54.130236 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9zm9z" Sep 30 14:26:54 crc kubenswrapper[4840]: I0930 14:26:54.137231 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xkx5r"] Sep 30 14:26:54 crc kubenswrapper[4840]: I0930 14:26:54.237052 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86b7d072-1a56-454f-ad5f-a098a0977122-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xkx5r\" (UID: \"86b7d072-1a56-454f-ad5f-a098a0977122\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xkx5r" Sep 30 14:26:54 crc kubenswrapper[4840]: I0930 14:26:54.237116 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcbbb\" (UniqueName: \"kubernetes.io/projected/86b7d072-1a56-454f-ad5f-a098a0977122-kube-api-access-dcbbb\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xkx5r\" (UID: \"86b7d072-1a56-454f-ad5f-a098a0977122\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xkx5r" Sep 30 14:26:54 crc kubenswrapper[4840]: I0930 14:26:54.237159 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86b7d072-1a56-454f-ad5f-a098a0977122-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xkx5r\" (UID: \"86b7d072-1a56-454f-ad5f-a098a0977122\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xkx5r" Sep 30 14:26:54 crc kubenswrapper[4840]: I0930 14:26:54.338649 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86b7d072-1a56-454f-ad5f-a098a0977122-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xkx5r\" (UID: \"86b7d072-1a56-454f-ad5f-a098a0977122\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xkx5r" Sep 30 14:26:54 crc kubenswrapper[4840]: I0930 14:26:54.338726 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcbbb\" (UniqueName: \"kubernetes.io/projected/86b7d072-1a56-454f-ad5f-a098a0977122-kube-api-access-dcbbb\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xkx5r\" (UID: \"86b7d072-1a56-454f-ad5f-a098a0977122\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xkx5r" Sep 30 14:26:54 crc kubenswrapper[4840]: I0930 14:26:54.338775 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86b7d072-1a56-454f-ad5f-a098a0977122-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xkx5r\" (UID: \"86b7d072-1a56-454f-ad5f-a098a0977122\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xkx5r" Sep 30 14:26:54 crc kubenswrapper[4840]: I0930 14:26:54.342921 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86b7d072-1a56-454f-ad5f-a098a0977122-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xkx5r\" (UID: \"86b7d072-1a56-454f-ad5f-a098a0977122\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xkx5r" Sep 30 14:26:54 crc kubenswrapper[4840]: I0930 14:26:54.344258 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86b7d072-1a56-454f-ad5f-a098a0977122-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xkx5r\" (UID: \"86b7d072-1a56-454f-ad5f-a098a0977122\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xkx5r" Sep 30 14:26:54 crc kubenswrapper[4840]: I0930 14:26:54.357377 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcbbb\" (UniqueName: \"kubernetes.io/projected/86b7d072-1a56-454f-ad5f-a098a0977122-kube-api-access-dcbbb\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xkx5r\" (UID: \"86b7d072-1a56-454f-ad5f-a098a0977122\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xkx5r" Sep 30 14:26:54 crc kubenswrapper[4840]: I0930 14:26:54.427787 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xkx5r" Sep 30 14:26:54 crc kubenswrapper[4840]: I0930 14:26:54.921937 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xkx5r"] Sep 30 14:26:55 crc kubenswrapper[4840]: I0930 14:26:55.042086 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xkx5r" event={"ID":"86b7d072-1a56-454f-ad5f-a098a0977122","Type":"ContainerStarted","Data":"8af2b940d6dafa6818724c3bc29af38b2a7ad407e9debc7255fa4671084a44e3"} Sep 30 14:26:56 crc kubenswrapper[4840]: I0930 14:26:56.049531 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xkx5r" event={"ID":"86b7d072-1a56-454f-ad5f-a098a0977122","Type":"ContainerStarted","Data":"b55d6005b4535e0bb39b052aed20c1db4d03ec434b999a98e2195e9e94fb66a5"} Sep 30 14:26:56 crc kubenswrapper[4840]: I0930 14:26:56.073220 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xkx5r" podStartSLOduration=1.6466141859999999 podStartE2EDuration="2.07320003s" podCreationTimestamp="2025-09-30 14:26:54 +0000 UTC" firstStartedPulling="2025-09-30 14:26:54.933128499 +0000 UTC m=+1843.562214912" lastFinishedPulling="2025-09-30 14:26:55.359714333 +0000 UTC m=+1843.988800756" observedRunningTime="2025-09-30 14:26:56.063213553 +0000 UTC m=+1844.692300006" watchObservedRunningTime="2025-09-30 14:26:56.07320003 +0000 UTC m=+1844.702286453" Sep 30 14:26:58 crc kubenswrapper[4840]: I0930 14:26:58.117308 4840 scope.go:117] "RemoveContainer" containerID="74b12cc9318ac99161f1b2541861d89b6100ba9a986e5863f53ae26c2a8953e7" Sep 30 14:26:58 crc kubenswrapper[4840]: E0930 14:26:58.117856 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:27:05 crc kubenswrapper[4840]: I0930 14:27:05.125830 4840 generic.go:334] "Generic (PLEG): container finished" podID="86b7d072-1a56-454f-ad5f-a098a0977122" containerID="b55d6005b4535e0bb39b052aed20c1db4d03ec434b999a98e2195e9e94fb66a5" exitCode=0 Sep 30 14:27:05 crc kubenswrapper[4840]: I0930 14:27:05.125907 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xkx5r" event={"ID":"86b7d072-1a56-454f-ad5f-a098a0977122","Type":"ContainerDied","Data":"b55d6005b4535e0bb39b052aed20c1db4d03ec434b999a98e2195e9e94fb66a5"} Sep 30 14:27:06 crc kubenswrapper[4840]: I0930 14:27:06.524216 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xkx5r" Sep 30 14:27:06 crc kubenswrapper[4840]: I0930 14:27:06.670816 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dcbbb\" (UniqueName: \"kubernetes.io/projected/86b7d072-1a56-454f-ad5f-a098a0977122-kube-api-access-dcbbb\") pod \"86b7d072-1a56-454f-ad5f-a098a0977122\" (UID: \"86b7d072-1a56-454f-ad5f-a098a0977122\") " Sep 30 14:27:06 crc kubenswrapper[4840]: I0930 14:27:06.671032 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86b7d072-1a56-454f-ad5f-a098a0977122-ssh-key\") pod \"86b7d072-1a56-454f-ad5f-a098a0977122\" (UID: \"86b7d072-1a56-454f-ad5f-a098a0977122\") " Sep 30 14:27:06 crc kubenswrapper[4840]: I0930 14:27:06.671090 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86b7d072-1a56-454f-ad5f-a098a0977122-inventory\") pod \"86b7d072-1a56-454f-ad5f-a098a0977122\" (UID: \"86b7d072-1a56-454f-ad5f-a098a0977122\") " Sep 30 14:27:06 crc kubenswrapper[4840]: I0930 14:27:06.677420 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86b7d072-1a56-454f-ad5f-a098a0977122-kube-api-access-dcbbb" (OuterVolumeSpecName: "kube-api-access-dcbbb") pod "86b7d072-1a56-454f-ad5f-a098a0977122" (UID: "86b7d072-1a56-454f-ad5f-a098a0977122"). InnerVolumeSpecName "kube-api-access-dcbbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:27:06 crc kubenswrapper[4840]: I0930 14:27:06.698931 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86b7d072-1a56-454f-ad5f-a098a0977122-inventory" (OuterVolumeSpecName: "inventory") pod "86b7d072-1a56-454f-ad5f-a098a0977122" (UID: "86b7d072-1a56-454f-ad5f-a098a0977122"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:27:06 crc kubenswrapper[4840]: I0930 14:27:06.699347 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86b7d072-1a56-454f-ad5f-a098a0977122-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "86b7d072-1a56-454f-ad5f-a098a0977122" (UID: "86b7d072-1a56-454f-ad5f-a098a0977122"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:27:06 crc kubenswrapper[4840]: I0930 14:27:06.773429 4840 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86b7d072-1a56-454f-ad5f-a098a0977122-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:27:06 crc kubenswrapper[4840]: I0930 14:27:06.773469 4840 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86b7d072-1a56-454f-ad5f-a098a0977122-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 14:27:06 crc kubenswrapper[4840]: I0930 14:27:06.773483 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dcbbb\" (UniqueName: \"kubernetes.io/projected/86b7d072-1a56-454f-ad5f-a098a0977122-kube-api-access-dcbbb\") on node \"crc\" DevicePath \"\"" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.164499 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xkx5r" event={"ID":"86b7d072-1a56-454f-ad5f-a098a0977122","Type":"ContainerDied","Data":"8af2b940d6dafa6818724c3bc29af38b2a7ad407e9debc7255fa4671084a44e3"} Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.164544 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8af2b940d6dafa6818724c3bc29af38b2a7ad407e9debc7255fa4671084a44e3" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.164639 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xkx5r" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.233171 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx"] Sep 30 14:27:07 crc kubenswrapper[4840]: E0930 14:27:07.233668 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86b7d072-1a56-454f-ad5f-a098a0977122" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.233688 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="86b7d072-1a56-454f-ad5f-a098a0977122" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.233907 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="86b7d072-1a56-454f-ad5f-a098a0977122" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.234565 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.238166 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.238271 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.238375 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.238492 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9zm9z" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.238502 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.238499 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.238777 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.240982 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.256833 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx"] Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.385098 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.385175 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cd627175-3b3b-4a69-9f1c-f64828267f40-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.385381 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.385445 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.385526 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.385582 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.385612 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cd627175-3b3b-4a69-9f1c-f64828267f40-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.385678 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.385727 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.385778 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cd627175-3b3b-4a69-9f1c-f64828267f40-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.385846 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cd627175-3b3b-4a69-9f1c-f64828267f40-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.385875 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.385896 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.385944 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7hqb\" (UniqueName: \"kubernetes.io/projected/cd627175-3b3b-4a69-9f1c-f64828267f40-kube-api-access-c7hqb\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.486910 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.486958 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cd627175-3b3b-4a69-9f1c-f64828267f40-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.487012 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.487037 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.487078 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.487096 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cd627175-3b3b-4a69-9f1c-f64828267f40-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.487114 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.487140 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.487164 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.487181 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cd627175-3b3b-4a69-9f1c-f64828267f40-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.487207 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cd627175-3b3b-4a69-9f1c-f64828267f40-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.487225 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.487243 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.487278 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7hqb\" (UniqueName: \"kubernetes.io/projected/cd627175-3b3b-4a69-9f1c-f64828267f40-kube-api-access-c7hqb\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.494828 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.494873 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cd627175-3b3b-4a69-9f1c-f64828267f40-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.495638 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.497035 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.499875 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cd627175-3b3b-4a69-9f1c-f64828267f40-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.500041 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.500052 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cd627175-3b3b-4a69-9f1c-f64828267f40-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.500180 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.500744 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.502383 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.502968 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cd627175-3b3b-4a69-9f1c-f64828267f40-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.505950 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.506600 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7hqb\" (UniqueName: \"kubernetes.io/projected/cd627175-3b3b-4a69-9f1c-f64828267f40-kube-api-access-c7hqb\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.506626 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:07 crc kubenswrapper[4840]: I0930 14:27:07.562882 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:08 crc kubenswrapper[4840]: I0930 14:27:08.074160 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx"] Sep 30 14:27:08 crc kubenswrapper[4840]: I0930 14:27:08.191319 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" event={"ID":"cd627175-3b3b-4a69-9f1c-f64828267f40","Type":"ContainerStarted","Data":"98e62984cd6f28ff719e47971115766e3241698fe6bdc66d37abd427dab57fdb"} Sep 30 14:27:09 crc kubenswrapper[4840]: I0930 14:27:09.219975 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" event={"ID":"cd627175-3b3b-4a69-9f1c-f64828267f40","Type":"ContainerStarted","Data":"d6e35a1dedc000640e46363a243065447c154896d7fd9d80be6e6f563fabc84f"} Sep 30 14:27:09 crc kubenswrapper[4840]: I0930 14:27:09.247482 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" podStartSLOduration=1.6359496949999999 podStartE2EDuration="2.247463061s" podCreationTimestamp="2025-09-30 14:27:07 +0000 UTC" firstStartedPulling="2025-09-30 14:27:08.078940599 +0000 UTC m=+1856.708027022" lastFinishedPulling="2025-09-30 14:27:08.690453965 +0000 UTC m=+1857.319540388" observedRunningTime="2025-09-30 14:27:09.245801508 +0000 UTC m=+1857.874887991" watchObservedRunningTime="2025-09-30 14:27:09.247463061 +0000 UTC m=+1857.876549484" Sep 30 14:27:13 crc kubenswrapper[4840]: I0930 14:27:13.116985 4840 scope.go:117] "RemoveContainer" containerID="74b12cc9318ac99161f1b2541861d89b6100ba9a986e5863f53ae26c2a8953e7" Sep 30 14:27:13 crc kubenswrapper[4840]: E0930 14:27:13.117746 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:27:24 crc kubenswrapper[4840]: I0930 14:27:24.117217 4840 scope.go:117] "RemoveContainer" containerID="74b12cc9318ac99161f1b2541861d89b6100ba9a986e5863f53ae26c2a8953e7" Sep 30 14:27:24 crc kubenswrapper[4840]: E0930 14:27:24.118093 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:27:38 crc kubenswrapper[4840]: I0930 14:27:38.116335 4840 scope.go:117] "RemoveContainer" containerID="74b12cc9318ac99161f1b2541861d89b6100ba9a986e5863f53ae26c2a8953e7" Sep 30 14:27:38 crc kubenswrapper[4840]: E0930 14:27:38.117085 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:27:45 crc kubenswrapper[4840]: I0930 14:27:45.531210 4840 generic.go:334] "Generic (PLEG): container finished" podID="cd627175-3b3b-4a69-9f1c-f64828267f40" containerID="d6e35a1dedc000640e46363a243065447c154896d7fd9d80be6e6f563fabc84f" exitCode=0 Sep 30 14:27:45 crc kubenswrapper[4840]: I0930 14:27:45.531284 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" event={"ID":"cd627175-3b3b-4a69-9f1c-f64828267f40","Type":"ContainerDied","Data":"d6e35a1dedc000640e46363a243065447c154896d7fd9d80be6e6f563fabc84f"} Sep 30 14:27:46 crc kubenswrapper[4840]: I0930 14:27:46.906848 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.031994 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7hqb\" (UniqueName: \"kubernetes.io/projected/cd627175-3b3b-4a69-9f1c-f64828267f40-kube-api-access-c7hqb\") pod \"cd627175-3b3b-4a69-9f1c-f64828267f40\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.032039 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-libvirt-combined-ca-bundle\") pod \"cd627175-3b3b-4a69-9f1c-f64828267f40\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.032079 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-nova-combined-ca-bundle\") pod \"cd627175-3b3b-4a69-9f1c-f64828267f40\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.032118 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-telemetry-combined-ca-bundle\") pod \"cd627175-3b3b-4a69-9f1c-f64828267f40\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.032145 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cd627175-3b3b-4a69-9f1c-f64828267f40-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"cd627175-3b3b-4a69-9f1c-f64828267f40\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.032184 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-neutron-metadata-combined-ca-bundle\") pod \"cd627175-3b3b-4a69-9f1c-f64828267f40\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.032225 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cd627175-3b3b-4a69-9f1c-f64828267f40-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"cd627175-3b3b-4a69-9f1c-f64828267f40\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.032260 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-repo-setup-combined-ca-bundle\") pod \"cd627175-3b3b-4a69-9f1c-f64828267f40\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.032329 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-inventory\") pod \"cd627175-3b3b-4a69-9f1c-f64828267f40\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.032401 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-ssh-key\") pod \"cd627175-3b3b-4a69-9f1c-f64828267f40\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.032426 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cd627175-3b3b-4a69-9f1c-f64828267f40-openstack-edpm-ipam-ovn-default-certs-0\") pod \"cd627175-3b3b-4a69-9f1c-f64828267f40\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.032464 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-bootstrap-combined-ca-bundle\") pod \"cd627175-3b3b-4a69-9f1c-f64828267f40\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.032493 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-ovn-combined-ca-bundle\") pod \"cd627175-3b3b-4a69-9f1c-f64828267f40\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.032527 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cd627175-3b3b-4a69-9f1c-f64828267f40-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"cd627175-3b3b-4a69-9f1c-f64828267f40\" (UID: \"cd627175-3b3b-4a69-9f1c-f64828267f40\") " Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.038450 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd627175-3b3b-4a69-9f1c-f64828267f40-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "cd627175-3b3b-4a69-9f1c-f64828267f40" (UID: "cd627175-3b3b-4a69-9f1c-f64828267f40"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.039006 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "cd627175-3b3b-4a69-9f1c-f64828267f40" (UID: "cd627175-3b3b-4a69-9f1c-f64828267f40"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.039364 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "cd627175-3b3b-4a69-9f1c-f64828267f40" (UID: "cd627175-3b3b-4a69-9f1c-f64828267f40"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.039420 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "cd627175-3b3b-4a69-9f1c-f64828267f40" (UID: "cd627175-3b3b-4a69-9f1c-f64828267f40"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.039458 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd627175-3b3b-4a69-9f1c-f64828267f40-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "cd627175-3b3b-4a69-9f1c-f64828267f40" (UID: "cd627175-3b3b-4a69-9f1c-f64828267f40"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.039494 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd627175-3b3b-4a69-9f1c-f64828267f40-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "cd627175-3b3b-4a69-9f1c-f64828267f40" (UID: "cd627175-3b3b-4a69-9f1c-f64828267f40"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.039982 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd627175-3b3b-4a69-9f1c-f64828267f40-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "cd627175-3b3b-4a69-9f1c-f64828267f40" (UID: "cd627175-3b3b-4a69-9f1c-f64828267f40"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.040487 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "cd627175-3b3b-4a69-9f1c-f64828267f40" (UID: "cd627175-3b3b-4a69-9f1c-f64828267f40"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.040736 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "cd627175-3b3b-4a69-9f1c-f64828267f40" (UID: "cd627175-3b3b-4a69-9f1c-f64828267f40"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.041618 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd627175-3b3b-4a69-9f1c-f64828267f40-kube-api-access-c7hqb" (OuterVolumeSpecName: "kube-api-access-c7hqb") pod "cd627175-3b3b-4a69-9f1c-f64828267f40" (UID: "cd627175-3b3b-4a69-9f1c-f64828267f40"). InnerVolumeSpecName "kube-api-access-c7hqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.041643 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "cd627175-3b3b-4a69-9f1c-f64828267f40" (UID: "cd627175-3b3b-4a69-9f1c-f64828267f40"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.042094 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "cd627175-3b3b-4a69-9f1c-f64828267f40" (UID: "cd627175-3b3b-4a69-9f1c-f64828267f40"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.065701 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-inventory" (OuterVolumeSpecName: "inventory") pod "cd627175-3b3b-4a69-9f1c-f64828267f40" (UID: "cd627175-3b3b-4a69-9f1c-f64828267f40"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.082421 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "cd627175-3b3b-4a69-9f1c-f64828267f40" (UID: "cd627175-3b3b-4a69-9f1c-f64828267f40"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.135094 4840 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.135139 4840 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cd627175-3b3b-4a69-9f1c-f64828267f40-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.135154 4840 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.135168 4840 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.135179 4840 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.135190 4840 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cd627175-3b3b-4a69-9f1c-f64828267f40-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.135205 4840 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.135218 4840 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.135229 4840 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cd627175-3b3b-4a69-9f1c-f64828267f40-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.135242 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7hqb\" (UniqueName: \"kubernetes.io/projected/cd627175-3b3b-4a69-9f1c-f64828267f40-kube-api-access-c7hqb\") on node \"crc\" DevicePath \"\"" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.135253 4840 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.135266 4840 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.135279 4840 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd627175-3b3b-4a69-9f1c-f64828267f40-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.135289 4840 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cd627175-3b3b-4a69-9f1c-f64828267f40-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.551306 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" event={"ID":"cd627175-3b3b-4a69-9f1c-f64828267f40","Type":"ContainerDied","Data":"98e62984cd6f28ff719e47971115766e3241698fe6bdc66d37abd427dab57fdb"} Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.551355 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="98e62984cd6f28ff719e47971115766e3241698fe6bdc66d37abd427dab57fdb" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.551457 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.637502 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-tsm4h"] Sep 30 14:27:47 crc kubenswrapper[4840]: E0930 14:27:47.637920 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd627175-3b3b-4a69-9f1c-f64828267f40" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.637934 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd627175-3b3b-4a69-9f1c-f64828267f40" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.638105 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd627175-3b3b-4a69-9f1c-f64828267f40" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.638782 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tsm4h" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.640766 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.640773 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9zm9z" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.641500 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.641714 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.644901 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.651834 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-tsm4h"] Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.747322 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/bc398435-386f-4d82-8ba5-618d3795223e-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tsm4h\" (UID: \"bc398435-386f-4d82-8ba5-618d3795223e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tsm4h" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.747421 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wjkq\" (UniqueName: \"kubernetes.io/projected/bc398435-386f-4d82-8ba5-618d3795223e-kube-api-access-4wjkq\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tsm4h\" (UID: \"bc398435-386f-4d82-8ba5-618d3795223e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tsm4h" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.747451 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc398435-386f-4d82-8ba5-618d3795223e-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tsm4h\" (UID: \"bc398435-386f-4d82-8ba5-618d3795223e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tsm4h" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.747512 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc398435-386f-4d82-8ba5-618d3795223e-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tsm4h\" (UID: \"bc398435-386f-4d82-8ba5-618d3795223e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tsm4h" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.747788 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc398435-386f-4d82-8ba5-618d3795223e-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tsm4h\" (UID: \"bc398435-386f-4d82-8ba5-618d3795223e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tsm4h" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.849739 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wjkq\" (UniqueName: \"kubernetes.io/projected/bc398435-386f-4d82-8ba5-618d3795223e-kube-api-access-4wjkq\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tsm4h\" (UID: \"bc398435-386f-4d82-8ba5-618d3795223e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tsm4h" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.849816 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc398435-386f-4d82-8ba5-618d3795223e-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tsm4h\" (UID: \"bc398435-386f-4d82-8ba5-618d3795223e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tsm4h" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.849903 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc398435-386f-4d82-8ba5-618d3795223e-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tsm4h\" (UID: \"bc398435-386f-4d82-8ba5-618d3795223e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tsm4h" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.849974 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc398435-386f-4d82-8ba5-618d3795223e-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tsm4h\" (UID: \"bc398435-386f-4d82-8ba5-618d3795223e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tsm4h" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.850066 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/bc398435-386f-4d82-8ba5-618d3795223e-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tsm4h\" (UID: \"bc398435-386f-4d82-8ba5-618d3795223e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tsm4h" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.851087 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/bc398435-386f-4d82-8ba5-618d3795223e-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tsm4h\" (UID: \"bc398435-386f-4d82-8ba5-618d3795223e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tsm4h" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.854438 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc398435-386f-4d82-8ba5-618d3795223e-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tsm4h\" (UID: \"bc398435-386f-4d82-8ba5-618d3795223e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tsm4h" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.854599 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc398435-386f-4d82-8ba5-618d3795223e-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tsm4h\" (UID: \"bc398435-386f-4d82-8ba5-618d3795223e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tsm4h" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.855203 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc398435-386f-4d82-8ba5-618d3795223e-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tsm4h\" (UID: \"bc398435-386f-4d82-8ba5-618d3795223e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tsm4h" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.866914 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wjkq\" (UniqueName: \"kubernetes.io/projected/bc398435-386f-4d82-8ba5-618d3795223e-kube-api-access-4wjkq\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tsm4h\" (UID: \"bc398435-386f-4d82-8ba5-618d3795223e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tsm4h" Sep 30 14:27:47 crc kubenswrapper[4840]: I0930 14:27:47.956043 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tsm4h" Sep 30 14:27:48 crc kubenswrapper[4840]: I0930 14:27:48.500616 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-tsm4h"] Sep 30 14:27:48 crc kubenswrapper[4840]: W0930 14:27:48.510147 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc398435_386f_4d82_8ba5_618d3795223e.slice/crio-b092134a634bbe5e47bdf9ee435e752af71f8353e94cdc3134cd0b449a790b87 WatchSource:0}: Error finding container b092134a634bbe5e47bdf9ee435e752af71f8353e94cdc3134cd0b449a790b87: Status 404 returned error can't find the container with id b092134a634bbe5e47bdf9ee435e752af71f8353e94cdc3134cd0b449a790b87 Sep 30 14:27:48 crc kubenswrapper[4840]: I0930 14:27:48.513769 4840 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 14:27:48 crc kubenswrapper[4840]: I0930 14:27:48.562045 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tsm4h" event={"ID":"bc398435-386f-4d82-8ba5-618d3795223e","Type":"ContainerStarted","Data":"b092134a634bbe5e47bdf9ee435e752af71f8353e94cdc3134cd0b449a790b87"} Sep 30 14:27:49 crc kubenswrapper[4840]: I0930 14:27:49.570581 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tsm4h" event={"ID":"bc398435-386f-4d82-8ba5-618d3795223e","Type":"ContainerStarted","Data":"1523bfc00702013e97ea5ebe4d12b9c63b08c3dfd55431ac0fec4463211e7458"} Sep 30 14:27:49 crc kubenswrapper[4840]: I0930 14:27:49.586014 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tsm4h" podStartSLOduration=2.15560595 podStartE2EDuration="2.585999048s" podCreationTimestamp="2025-09-30 14:27:47 +0000 UTC" firstStartedPulling="2025-09-30 14:27:48.513560073 +0000 UTC m=+1897.142646486" lastFinishedPulling="2025-09-30 14:27:48.943953161 +0000 UTC m=+1897.573039584" observedRunningTime="2025-09-30 14:27:49.583751591 +0000 UTC m=+1898.212838044" watchObservedRunningTime="2025-09-30 14:27:49.585999048 +0000 UTC m=+1898.215085471" Sep 30 14:27:52 crc kubenswrapper[4840]: I0930 14:27:52.122379 4840 scope.go:117] "RemoveContainer" containerID="74b12cc9318ac99161f1b2541861d89b6100ba9a986e5863f53ae26c2a8953e7" Sep 30 14:27:52 crc kubenswrapper[4840]: E0930 14:27:52.123231 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:28:07 crc kubenswrapper[4840]: I0930 14:28:07.116631 4840 scope.go:117] "RemoveContainer" containerID="74b12cc9318ac99161f1b2541861d89b6100ba9a986e5863f53ae26c2a8953e7" Sep 30 14:28:07 crc kubenswrapper[4840]: E0930 14:28:07.117425 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:28:22 crc kubenswrapper[4840]: I0930 14:28:22.121856 4840 scope.go:117] "RemoveContainer" containerID="74b12cc9318ac99161f1b2541861d89b6100ba9a986e5863f53ae26c2a8953e7" Sep 30 14:28:22 crc kubenswrapper[4840]: E0930 14:28:22.122626 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:28:33 crc kubenswrapper[4840]: I0930 14:28:33.115947 4840 scope.go:117] "RemoveContainer" containerID="74b12cc9318ac99161f1b2541861d89b6100ba9a986e5863f53ae26c2a8953e7" Sep 30 14:28:33 crc kubenswrapper[4840]: E0930 14:28:33.116697 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:28:46 crc kubenswrapper[4840]: I0930 14:28:46.117067 4840 scope.go:117] "RemoveContainer" containerID="74b12cc9318ac99161f1b2541861d89b6100ba9a986e5863f53ae26c2a8953e7" Sep 30 14:28:46 crc kubenswrapper[4840]: E0930 14:28:46.117889 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:28:51 crc kubenswrapper[4840]: I0930 14:28:51.101674 4840 generic.go:334] "Generic (PLEG): container finished" podID="bc398435-386f-4d82-8ba5-618d3795223e" containerID="1523bfc00702013e97ea5ebe4d12b9c63b08c3dfd55431ac0fec4463211e7458" exitCode=0 Sep 30 14:28:51 crc kubenswrapper[4840]: I0930 14:28:51.101759 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tsm4h" event={"ID":"bc398435-386f-4d82-8ba5-618d3795223e","Type":"ContainerDied","Data":"1523bfc00702013e97ea5ebe4d12b9c63b08c3dfd55431ac0fec4463211e7458"} Sep 30 14:28:52 crc kubenswrapper[4840]: I0930 14:28:52.517777 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tsm4h" Sep 30 14:28:52 crc kubenswrapper[4840]: I0930 14:28:52.627150 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc398435-386f-4d82-8ba5-618d3795223e-ssh-key\") pod \"bc398435-386f-4d82-8ba5-618d3795223e\" (UID: \"bc398435-386f-4d82-8ba5-618d3795223e\") " Sep 30 14:28:52 crc kubenswrapper[4840]: I0930 14:28:52.627205 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc398435-386f-4d82-8ba5-618d3795223e-ovn-combined-ca-bundle\") pod \"bc398435-386f-4d82-8ba5-618d3795223e\" (UID: \"bc398435-386f-4d82-8ba5-618d3795223e\") " Sep 30 14:28:52 crc kubenswrapper[4840]: I0930 14:28:52.627341 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/bc398435-386f-4d82-8ba5-618d3795223e-ovncontroller-config-0\") pod \"bc398435-386f-4d82-8ba5-618d3795223e\" (UID: \"bc398435-386f-4d82-8ba5-618d3795223e\") " Sep 30 14:28:52 crc kubenswrapper[4840]: I0930 14:28:52.627379 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc398435-386f-4d82-8ba5-618d3795223e-inventory\") pod \"bc398435-386f-4d82-8ba5-618d3795223e\" (UID: \"bc398435-386f-4d82-8ba5-618d3795223e\") " Sep 30 14:28:52 crc kubenswrapper[4840]: I0930 14:28:52.627434 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wjkq\" (UniqueName: \"kubernetes.io/projected/bc398435-386f-4d82-8ba5-618d3795223e-kube-api-access-4wjkq\") pod \"bc398435-386f-4d82-8ba5-618d3795223e\" (UID: \"bc398435-386f-4d82-8ba5-618d3795223e\") " Sep 30 14:28:52 crc kubenswrapper[4840]: I0930 14:28:52.633046 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc398435-386f-4d82-8ba5-618d3795223e-kube-api-access-4wjkq" (OuterVolumeSpecName: "kube-api-access-4wjkq") pod "bc398435-386f-4d82-8ba5-618d3795223e" (UID: "bc398435-386f-4d82-8ba5-618d3795223e"). InnerVolumeSpecName "kube-api-access-4wjkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:28:52 crc kubenswrapper[4840]: I0930 14:28:52.635883 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc398435-386f-4d82-8ba5-618d3795223e-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "bc398435-386f-4d82-8ba5-618d3795223e" (UID: "bc398435-386f-4d82-8ba5-618d3795223e"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:28:52 crc kubenswrapper[4840]: I0930 14:28:52.655988 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc398435-386f-4d82-8ba5-618d3795223e-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "bc398435-386f-4d82-8ba5-618d3795223e" (UID: "bc398435-386f-4d82-8ba5-618d3795223e"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:28:52 crc kubenswrapper[4840]: I0930 14:28:52.664214 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc398435-386f-4d82-8ba5-618d3795223e-inventory" (OuterVolumeSpecName: "inventory") pod "bc398435-386f-4d82-8ba5-618d3795223e" (UID: "bc398435-386f-4d82-8ba5-618d3795223e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:28:52 crc kubenswrapper[4840]: I0930 14:28:52.664827 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc398435-386f-4d82-8ba5-618d3795223e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bc398435-386f-4d82-8ba5-618d3795223e" (UID: "bc398435-386f-4d82-8ba5-618d3795223e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:28:52 crc kubenswrapper[4840]: I0930 14:28:52.729594 4840 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc398435-386f-4d82-8ba5-618d3795223e-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:28:52 crc kubenswrapper[4840]: I0930 14:28:52.729624 4840 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc398435-386f-4d82-8ba5-618d3795223e-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:28:52 crc kubenswrapper[4840]: I0930 14:28:52.729636 4840 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/bc398435-386f-4d82-8ba5-618d3795223e-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:28:52 crc kubenswrapper[4840]: I0930 14:28:52.729644 4840 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc398435-386f-4d82-8ba5-618d3795223e-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 14:28:52 crc kubenswrapper[4840]: I0930 14:28:52.729653 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wjkq\" (UniqueName: \"kubernetes.io/projected/bc398435-386f-4d82-8ba5-618d3795223e-kube-api-access-4wjkq\") on node \"crc\" DevicePath \"\"" Sep 30 14:28:53 crc kubenswrapper[4840]: I0930 14:28:53.119998 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tsm4h" event={"ID":"bc398435-386f-4d82-8ba5-618d3795223e","Type":"ContainerDied","Data":"b092134a634bbe5e47bdf9ee435e752af71f8353e94cdc3134cd0b449a790b87"} Sep 30 14:28:53 crc kubenswrapper[4840]: I0930 14:28:53.120041 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b092134a634bbe5e47bdf9ee435e752af71f8353e94cdc3134cd0b449a790b87" Sep 30 14:28:53 crc kubenswrapper[4840]: I0930 14:28:53.120089 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tsm4h" Sep 30 14:28:53 crc kubenswrapper[4840]: I0930 14:28:53.200069 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8"] Sep 30 14:28:53 crc kubenswrapper[4840]: E0930 14:28:53.200779 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc398435-386f-4d82-8ba5-618d3795223e" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Sep 30 14:28:53 crc kubenswrapper[4840]: I0930 14:28:53.200799 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc398435-386f-4d82-8ba5-618d3795223e" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Sep 30 14:28:53 crc kubenswrapper[4840]: I0930 14:28:53.201023 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc398435-386f-4d82-8ba5-618d3795223e" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Sep 30 14:28:53 crc kubenswrapper[4840]: I0930 14:28:53.201706 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8" Sep 30 14:28:53 crc kubenswrapper[4840]: I0930 14:28:53.206921 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 14:28:53 crc kubenswrapper[4840]: I0930 14:28:53.207034 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 14:28:53 crc kubenswrapper[4840]: I0930 14:28:53.207181 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Sep 30 14:28:53 crc kubenswrapper[4840]: I0930 14:28:53.207332 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 14:28:53 crc kubenswrapper[4840]: I0930 14:28:53.207652 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Sep 30 14:28:53 crc kubenswrapper[4840]: I0930 14:28:53.209220 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9zm9z" Sep 30 14:28:53 crc kubenswrapper[4840]: I0930 14:28:53.219608 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8"] Sep 30 14:28:53 crc kubenswrapper[4840]: I0930 14:28:53.345601 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/800dcd1d-c19d-4841-b3ab-81c312d8a754-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8\" (UID: \"800dcd1d-c19d-4841-b3ab-81c312d8a754\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8" Sep 30 14:28:53 crc kubenswrapper[4840]: I0930 14:28:53.345673 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/800dcd1d-c19d-4841-b3ab-81c312d8a754-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8\" (UID: \"800dcd1d-c19d-4841-b3ab-81c312d8a754\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8" Sep 30 14:28:53 crc kubenswrapper[4840]: I0930 14:28:53.345705 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/800dcd1d-c19d-4841-b3ab-81c312d8a754-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8\" (UID: \"800dcd1d-c19d-4841-b3ab-81c312d8a754\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8" Sep 30 14:28:53 crc kubenswrapper[4840]: I0930 14:28:53.345744 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/800dcd1d-c19d-4841-b3ab-81c312d8a754-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8\" (UID: \"800dcd1d-c19d-4841-b3ab-81c312d8a754\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8" Sep 30 14:28:53 crc kubenswrapper[4840]: I0930 14:28:53.345775 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/800dcd1d-c19d-4841-b3ab-81c312d8a754-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8\" (UID: \"800dcd1d-c19d-4841-b3ab-81c312d8a754\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8" Sep 30 14:28:53 crc kubenswrapper[4840]: I0930 14:28:53.345840 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6x4wr\" (UniqueName: \"kubernetes.io/projected/800dcd1d-c19d-4841-b3ab-81c312d8a754-kube-api-access-6x4wr\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8\" (UID: \"800dcd1d-c19d-4841-b3ab-81c312d8a754\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8" Sep 30 14:28:53 crc kubenswrapper[4840]: I0930 14:28:53.447302 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/800dcd1d-c19d-4841-b3ab-81c312d8a754-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8\" (UID: \"800dcd1d-c19d-4841-b3ab-81c312d8a754\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8" Sep 30 14:28:53 crc kubenswrapper[4840]: I0930 14:28:53.447374 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/800dcd1d-c19d-4841-b3ab-81c312d8a754-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8\" (UID: \"800dcd1d-c19d-4841-b3ab-81c312d8a754\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8" Sep 30 14:28:53 crc kubenswrapper[4840]: I0930 14:28:53.447433 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6x4wr\" (UniqueName: \"kubernetes.io/projected/800dcd1d-c19d-4841-b3ab-81c312d8a754-kube-api-access-6x4wr\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8\" (UID: \"800dcd1d-c19d-4841-b3ab-81c312d8a754\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8" Sep 30 14:28:53 crc kubenswrapper[4840]: I0930 14:28:53.447567 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/800dcd1d-c19d-4841-b3ab-81c312d8a754-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8\" (UID: \"800dcd1d-c19d-4841-b3ab-81c312d8a754\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8" Sep 30 14:28:53 crc kubenswrapper[4840]: I0930 14:28:53.447633 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/800dcd1d-c19d-4841-b3ab-81c312d8a754-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8\" (UID: \"800dcd1d-c19d-4841-b3ab-81c312d8a754\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8" Sep 30 14:28:53 crc kubenswrapper[4840]: I0930 14:28:53.447663 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/800dcd1d-c19d-4841-b3ab-81c312d8a754-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8\" (UID: \"800dcd1d-c19d-4841-b3ab-81c312d8a754\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8" Sep 30 14:28:53 crc kubenswrapper[4840]: I0930 14:28:53.452797 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/800dcd1d-c19d-4841-b3ab-81c312d8a754-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8\" (UID: \"800dcd1d-c19d-4841-b3ab-81c312d8a754\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8" Sep 30 14:28:53 crc kubenswrapper[4840]: I0930 14:28:53.452913 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/800dcd1d-c19d-4841-b3ab-81c312d8a754-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8\" (UID: \"800dcd1d-c19d-4841-b3ab-81c312d8a754\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8" Sep 30 14:28:53 crc kubenswrapper[4840]: I0930 14:28:53.452982 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/800dcd1d-c19d-4841-b3ab-81c312d8a754-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8\" (UID: \"800dcd1d-c19d-4841-b3ab-81c312d8a754\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8" Sep 30 14:28:53 crc kubenswrapper[4840]: I0930 14:28:53.455203 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/800dcd1d-c19d-4841-b3ab-81c312d8a754-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8\" (UID: \"800dcd1d-c19d-4841-b3ab-81c312d8a754\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8" Sep 30 14:28:53 crc kubenswrapper[4840]: I0930 14:28:53.458036 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/800dcd1d-c19d-4841-b3ab-81c312d8a754-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8\" (UID: \"800dcd1d-c19d-4841-b3ab-81c312d8a754\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8" Sep 30 14:28:53 crc kubenswrapper[4840]: I0930 14:28:53.464279 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6x4wr\" (UniqueName: \"kubernetes.io/projected/800dcd1d-c19d-4841-b3ab-81c312d8a754-kube-api-access-6x4wr\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8\" (UID: \"800dcd1d-c19d-4841-b3ab-81c312d8a754\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8" Sep 30 14:28:53 crc kubenswrapper[4840]: I0930 14:28:53.531857 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8" Sep 30 14:28:54 crc kubenswrapper[4840]: I0930 14:28:54.054971 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8"] Sep 30 14:28:54 crc kubenswrapper[4840]: I0930 14:28:54.128719 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8" event={"ID":"800dcd1d-c19d-4841-b3ab-81c312d8a754","Type":"ContainerStarted","Data":"0538964629b5ce8fac3302b7f26204298dafec9c82ff04964d62a1dc385c16b8"} Sep 30 14:28:55 crc kubenswrapper[4840]: I0930 14:28:55.142842 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8" event={"ID":"800dcd1d-c19d-4841-b3ab-81c312d8a754","Type":"ContainerStarted","Data":"36bd63d199b48c19f354414a7038b7dc957f06b04d0b8f99ab8ebac90458d7d4"} Sep 30 14:28:57 crc kubenswrapper[4840]: I0930 14:28:57.117007 4840 scope.go:117] "RemoveContainer" containerID="74b12cc9318ac99161f1b2541861d89b6100ba9a986e5863f53ae26c2a8953e7" Sep 30 14:28:57 crc kubenswrapper[4840]: E0930 14:28:57.117260 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:29:09 crc kubenswrapper[4840]: I0930 14:29:09.116695 4840 scope.go:117] "RemoveContainer" containerID="74b12cc9318ac99161f1b2541861d89b6100ba9a986e5863f53ae26c2a8953e7" Sep 30 14:29:09 crc kubenswrapper[4840]: E0930 14:29:09.117592 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:29:21 crc kubenswrapper[4840]: I0930 14:29:21.116588 4840 scope.go:117] "RemoveContainer" containerID="74b12cc9318ac99161f1b2541861d89b6100ba9a986e5863f53ae26c2a8953e7" Sep 30 14:29:21 crc kubenswrapper[4840]: E0930 14:29:21.117442 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:29:33 crc kubenswrapper[4840]: I0930 14:29:33.116088 4840 scope.go:117] "RemoveContainer" containerID="74b12cc9318ac99161f1b2541861d89b6100ba9a986e5863f53ae26c2a8953e7" Sep 30 14:29:33 crc kubenswrapper[4840]: I0930 14:29:33.497035 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" event={"ID":"10e8b890-7f20-4a36-8e03-898620cf599a","Type":"ContainerStarted","Data":"6539ed77e1c5a86ff8a57fda399610c4bb26dcd9ac75c7c7b74f7ff71d0bfc80"} Sep 30 14:29:33 crc kubenswrapper[4840]: I0930 14:29:33.518015 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8" podStartSLOduration=39.925127332 podStartE2EDuration="40.517993199s" podCreationTimestamp="2025-09-30 14:28:53 +0000 UTC" firstStartedPulling="2025-09-30 14:28:54.058605632 +0000 UTC m=+1962.687692055" lastFinishedPulling="2025-09-30 14:28:54.651471499 +0000 UTC m=+1963.280557922" observedRunningTime="2025-09-30 14:28:55.162426029 +0000 UTC m=+1963.791512452" watchObservedRunningTime="2025-09-30 14:29:33.517993199 +0000 UTC m=+2002.147079622" Sep 30 14:29:41 crc kubenswrapper[4840]: I0930 14:29:41.562021 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8" event={"ID":"800dcd1d-c19d-4841-b3ab-81c312d8a754","Type":"ContainerDied","Data":"36bd63d199b48c19f354414a7038b7dc957f06b04d0b8f99ab8ebac90458d7d4"} Sep 30 14:29:41 crc kubenswrapper[4840]: I0930 14:29:41.561963 4840 generic.go:334] "Generic (PLEG): container finished" podID="800dcd1d-c19d-4841-b3ab-81c312d8a754" containerID="36bd63d199b48c19f354414a7038b7dc957f06b04d0b8f99ab8ebac90458d7d4" exitCode=0 Sep 30 14:29:43 crc kubenswrapper[4840]: I0930 14:29:43.060997 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8" Sep 30 14:29:43 crc kubenswrapper[4840]: I0930 14:29:43.209323 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6x4wr\" (UniqueName: \"kubernetes.io/projected/800dcd1d-c19d-4841-b3ab-81c312d8a754-kube-api-access-6x4wr\") pod \"800dcd1d-c19d-4841-b3ab-81c312d8a754\" (UID: \"800dcd1d-c19d-4841-b3ab-81c312d8a754\") " Sep 30 14:29:43 crc kubenswrapper[4840]: I0930 14:29:43.209416 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/800dcd1d-c19d-4841-b3ab-81c312d8a754-neutron-ovn-metadata-agent-neutron-config-0\") pod \"800dcd1d-c19d-4841-b3ab-81c312d8a754\" (UID: \"800dcd1d-c19d-4841-b3ab-81c312d8a754\") " Sep 30 14:29:43 crc kubenswrapper[4840]: I0930 14:29:43.209444 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/800dcd1d-c19d-4841-b3ab-81c312d8a754-inventory\") pod \"800dcd1d-c19d-4841-b3ab-81c312d8a754\" (UID: \"800dcd1d-c19d-4841-b3ab-81c312d8a754\") " Sep 30 14:29:43 crc kubenswrapper[4840]: I0930 14:29:43.209591 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/800dcd1d-c19d-4841-b3ab-81c312d8a754-ssh-key\") pod \"800dcd1d-c19d-4841-b3ab-81c312d8a754\" (UID: \"800dcd1d-c19d-4841-b3ab-81c312d8a754\") " Sep 30 14:29:43 crc kubenswrapper[4840]: I0930 14:29:43.209676 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/800dcd1d-c19d-4841-b3ab-81c312d8a754-neutron-metadata-combined-ca-bundle\") pod \"800dcd1d-c19d-4841-b3ab-81c312d8a754\" (UID: \"800dcd1d-c19d-4841-b3ab-81c312d8a754\") " Sep 30 14:29:43 crc kubenswrapper[4840]: I0930 14:29:43.209733 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/800dcd1d-c19d-4841-b3ab-81c312d8a754-nova-metadata-neutron-config-0\") pod \"800dcd1d-c19d-4841-b3ab-81c312d8a754\" (UID: \"800dcd1d-c19d-4841-b3ab-81c312d8a754\") " Sep 30 14:29:43 crc kubenswrapper[4840]: I0930 14:29:43.216140 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/800dcd1d-c19d-4841-b3ab-81c312d8a754-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "800dcd1d-c19d-4841-b3ab-81c312d8a754" (UID: "800dcd1d-c19d-4841-b3ab-81c312d8a754"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:29:43 crc kubenswrapper[4840]: I0930 14:29:43.216652 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/800dcd1d-c19d-4841-b3ab-81c312d8a754-kube-api-access-6x4wr" (OuterVolumeSpecName: "kube-api-access-6x4wr") pod "800dcd1d-c19d-4841-b3ab-81c312d8a754" (UID: "800dcd1d-c19d-4841-b3ab-81c312d8a754"). InnerVolumeSpecName "kube-api-access-6x4wr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:29:43 crc kubenswrapper[4840]: I0930 14:29:43.237426 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/800dcd1d-c19d-4841-b3ab-81c312d8a754-inventory" (OuterVolumeSpecName: "inventory") pod "800dcd1d-c19d-4841-b3ab-81c312d8a754" (UID: "800dcd1d-c19d-4841-b3ab-81c312d8a754"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:29:43 crc kubenswrapper[4840]: I0930 14:29:43.242765 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/800dcd1d-c19d-4841-b3ab-81c312d8a754-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "800dcd1d-c19d-4841-b3ab-81c312d8a754" (UID: "800dcd1d-c19d-4841-b3ab-81c312d8a754"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:29:43 crc kubenswrapper[4840]: I0930 14:29:43.245982 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/800dcd1d-c19d-4841-b3ab-81c312d8a754-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "800dcd1d-c19d-4841-b3ab-81c312d8a754" (UID: "800dcd1d-c19d-4841-b3ab-81c312d8a754"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:29:43 crc kubenswrapper[4840]: I0930 14:29:43.254949 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/800dcd1d-c19d-4841-b3ab-81c312d8a754-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "800dcd1d-c19d-4841-b3ab-81c312d8a754" (UID: "800dcd1d-c19d-4841-b3ab-81c312d8a754"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:29:43 crc kubenswrapper[4840]: I0930 14:29:43.311863 4840 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/800dcd1d-c19d-4841-b3ab-81c312d8a754-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:29:43 crc kubenswrapper[4840]: I0930 14:29:43.311894 4840 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/800dcd1d-c19d-4841-b3ab-81c312d8a754-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:29:43 crc kubenswrapper[4840]: I0930 14:29:43.311906 4840 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/800dcd1d-c19d-4841-b3ab-81c312d8a754-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:29:43 crc kubenswrapper[4840]: I0930 14:29:43.311915 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6x4wr\" (UniqueName: \"kubernetes.io/projected/800dcd1d-c19d-4841-b3ab-81c312d8a754-kube-api-access-6x4wr\") on node \"crc\" DevicePath \"\"" Sep 30 14:29:43 crc kubenswrapper[4840]: I0930 14:29:43.311925 4840 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/800dcd1d-c19d-4841-b3ab-81c312d8a754-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:29:43 crc kubenswrapper[4840]: I0930 14:29:43.311936 4840 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/800dcd1d-c19d-4841-b3ab-81c312d8a754-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 14:29:43 crc kubenswrapper[4840]: I0930 14:29:43.580255 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8" event={"ID":"800dcd1d-c19d-4841-b3ab-81c312d8a754","Type":"ContainerDied","Data":"0538964629b5ce8fac3302b7f26204298dafec9c82ff04964d62a1dc385c16b8"} Sep 30 14:29:43 crc kubenswrapper[4840]: I0930 14:29:43.580313 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0538964629b5ce8fac3302b7f26204298dafec9c82ff04964d62a1dc385c16b8" Sep 30 14:29:43 crc kubenswrapper[4840]: I0930 14:29:43.580324 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8" Sep 30 14:29:43 crc kubenswrapper[4840]: I0930 14:29:43.729836 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m584b"] Sep 30 14:29:43 crc kubenswrapper[4840]: E0930 14:29:43.730197 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="800dcd1d-c19d-4841-b3ab-81c312d8a754" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Sep 30 14:29:43 crc kubenswrapper[4840]: I0930 14:29:43.730218 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="800dcd1d-c19d-4841-b3ab-81c312d8a754" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Sep 30 14:29:43 crc kubenswrapper[4840]: I0930 14:29:43.730672 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="800dcd1d-c19d-4841-b3ab-81c312d8a754" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Sep 30 14:29:43 crc kubenswrapper[4840]: I0930 14:29:43.731361 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m584b" Sep 30 14:29:43 crc kubenswrapper[4840]: I0930 14:29:43.733396 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 14:29:43 crc kubenswrapper[4840]: I0930 14:29:43.734509 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 14:29:43 crc kubenswrapper[4840]: I0930 14:29:43.734758 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9zm9z" Sep 30 14:29:43 crc kubenswrapper[4840]: I0930 14:29:43.734856 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Sep 30 14:29:43 crc kubenswrapper[4840]: I0930 14:29:43.737681 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 14:29:43 crc kubenswrapper[4840]: I0930 14:29:43.742926 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m584b"] Sep 30 14:29:43 crc kubenswrapper[4840]: I0930 14:29:43.929690 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f66833a-f0de-410b-9459-80203d4291a6-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m584b\" (UID: \"9f66833a-f0de-410b-9459-80203d4291a6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m584b" Sep 30 14:29:43 crc kubenswrapper[4840]: I0930 14:29:43.929754 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/9f66833a-f0de-410b-9459-80203d4291a6-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m584b\" (UID: \"9f66833a-f0de-410b-9459-80203d4291a6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m584b" Sep 30 14:29:43 crc kubenswrapper[4840]: I0930 14:29:43.929808 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f66833a-f0de-410b-9459-80203d4291a6-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m584b\" (UID: \"9f66833a-f0de-410b-9459-80203d4291a6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m584b" Sep 30 14:29:43 crc kubenswrapper[4840]: I0930 14:29:43.930674 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f66833a-f0de-410b-9459-80203d4291a6-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m584b\" (UID: \"9f66833a-f0de-410b-9459-80203d4291a6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m584b" Sep 30 14:29:43 crc kubenswrapper[4840]: I0930 14:29:43.930798 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89j6f\" (UniqueName: \"kubernetes.io/projected/9f66833a-f0de-410b-9459-80203d4291a6-kube-api-access-89j6f\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m584b\" (UID: \"9f66833a-f0de-410b-9459-80203d4291a6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m584b" Sep 30 14:29:44 crc kubenswrapper[4840]: I0930 14:29:44.032998 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f66833a-f0de-410b-9459-80203d4291a6-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m584b\" (UID: \"9f66833a-f0de-410b-9459-80203d4291a6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m584b" Sep 30 14:29:44 crc kubenswrapper[4840]: I0930 14:29:44.033070 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89j6f\" (UniqueName: \"kubernetes.io/projected/9f66833a-f0de-410b-9459-80203d4291a6-kube-api-access-89j6f\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m584b\" (UID: \"9f66833a-f0de-410b-9459-80203d4291a6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m584b" Sep 30 14:29:44 crc kubenswrapper[4840]: I0930 14:29:44.033200 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f66833a-f0de-410b-9459-80203d4291a6-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m584b\" (UID: \"9f66833a-f0de-410b-9459-80203d4291a6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m584b" Sep 30 14:29:44 crc kubenswrapper[4840]: I0930 14:29:44.033230 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/9f66833a-f0de-410b-9459-80203d4291a6-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m584b\" (UID: \"9f66833a-f0de-410b-9459-80203d4291a6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m584b" Sep 30 14:29:44 crc kubenswrapper[4840]: I0930 14:29:44.033267 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f66833a-f0de-410b-9459-80203d4291a6-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m584b\" (UID: \"9f66833a-f0de-410b-9459-80203d4291a6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m584b" Sep 30 14:29:44 crc kubenswrapper[4840]: I0930 14:29:44.036972 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f66833a-f0de-410b-9459-80203d4291a6-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m584b\" (UID: \"9f66833a-f0de-410b-9459-80203d4291a6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m584b" Sep 30 14:29:44 crc kubenswrapper[4840]: I0930 14:29:44.037359 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f66833a-f0de-410b-9459-80203d4291a6-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m584b\" (UID: \"9f66833a-f0de-410b-9459-80203d4291a6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m584b" Sep 30 14:29:44 crc kubenswrapper[4840]: I0930 14:29:44.037628 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/9f66833a-f0de-410b-9459-80203d4291a6-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m584b\" (UID: \"9f66833a-f0de-410b-9459-80203d4291a6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m584b" Sep 30 14:29:44 crc kubenswrapper[4840]: I0930 14:29:44.038730 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f66833a-f0de-410b-9459-80203d4291a6-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m584b\" (UID: \"9f66833a-f0de-410b-9459-80203d4291a6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m584b" Sep 30 14:29:44 crc kubenswrapper[4840]: I0930 14:29:44.050531 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89j6f\" (UniqueName: \"kubernetes.io/projected/9f66833a-f0de-410b-9459-80203d4291a6-kube-api-access-89j6f\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m584b\" (UID: \"9f66833a-f0de-410b-9459-80203d4291a6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m584b" Sep 30 14:29:44 crc kubenswrapper[4840]: I0930 14:29:44.050953 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m584b" Sep 30 14:29:44 crc kubenswrapper[4840]: I0930 14:29:44.574150 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m584b"] Sep 30 14:29:44 crc kubenswrapper[4840]: W0930 14:29:44.576993 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f66833a_f0de_410b_9459_80203d4291a6.slice/crio-701804470c4d4cca4a5a0da1a8b9c7a3808f8a37e56140dc1807f7029d12bbb2 WatchSource:0}: Error finding container 701804470c4d4cca4a5a0da1a8b9c7a3808f8a37e56140dc1807f7029d12bbb2: Status 404 returned error can't find the container with id 701804470c4d4cca4a5a0da1a8b9c7a3808f8a37e56140dc1807f7029d12bbb2 Sep 30 14:29:44 crc kubenswrapper[4840]: I0930 14:29:44.590769 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m584b" event={"ID":"9f66833a-f0de-410b-9459-80203d4291a6","Type":"ContainerStarted","Data":"701804470c4d4cca4a5a0da1a8b9c7a3808f8a37e56140dc1807f7029d12bbb2"} Sep 30 14:29:46 crc kubenswrapper[4840]: I0930 14:29:46.612115 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m584b" event={"ID":"9f66833a-f0de-410b-9459-80203d4291a6","Type":"ContainerStarted","Data":"64a9bd6bceffc16ee4b6db867c83bf1cb58cac34b2aab3d521ee942e2334fd1a"} Sep 30 14:29:46 crc kubenswrapper[4840]: I0930 14:29:46.638027 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m584b" podStartSLOduration=2.402654814 podStartE2EDuration="3.637969985s" podCreationTimestamp="2025-09-30 14:29:43 +0000 UTC" firstStartedPulling="2025-09-30 14:29:44.578748707 +0000 UTC m=+2013.207835130" lastFinishedPulling="2025-09-30 14:29:45.814063868 +0000 UTC m=+2014.443150301" observedRunningTime="2025-09-30 14:29:46.63032738 +0000 UTC m=+2015.259413803" watchObservedRunningTime="2025-09-30 14:29:46.637969985 +0000 UTC m=+2015.267056408" Sep 30 14:30:00 crc kubenswrapper[4840]: I0930 14:30:00.154104 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320710-vtk5c"] Sep 30 14:30:00 crc kubenswrapper[4840]: I0930 14:30:00.156096 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320710-vtk5c" Sep 30 14:30:00 crc kubenswrapper[4840]: I0930 14:30:00.158157 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 30 14:30:00 crc kubenswrapper[4840]: I0930 14:30:00.164957 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 30 14:30:00 crc kubenswrapper[4840]: I0930 14:30:00.165668 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320710-vtk5c"] Sep 30 14:30:00 crc kubenswrapper[4840]: I0930 14:30:00.230482 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b27f0395-be2f-431c-baac-5eaeef3ab6c2-secret-volume\") pod \"collect-profiles-29320710-vtk5c\" (UID: \"b27f0395-be2f-431c-baac-5eaeef3ab6c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320710-vtk5c" Sep 30 14:30:00 crc kubenswrapper[4840]: I0930 14:30:00.230591 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7glv\" (UniqueName: \"kubernetes.io/projected/b27f0395-be2f-431c-baac-5eaeef3ab6c2-kube-api-access-s7glv\") pod \"collect-profiles-29320710-vtk5c\" (UID: \"b27f0395-be2f-431c-baac-5eaeef3ab6c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320710-vtk5c" Sep 30 14:30:00 crc kubenswrapper[4840]: I0930 14:30:00.230649 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b27f0395-be2f-431c-baac-5eaeef3ab6c2-config-volume\") pod \"collect-profiles-29320710-vtk5c\" (UID: \"b27f0395-be2f-431c-baac-5eaeef3ab6c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320710-vtk5c" Sep 30 14:30:00 crc kubenswrapper[4840]: I0930 14:30:00.332505 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b27f0395-be2f-431c-baac-5eaeef3ab6c2-secret-volume\") pod \"collect-profiles-29320710-vtk5c\" (UID: \"b27f0395-be2f-431c-baac-5eaeef3ab6c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320710-vtk5c" Sep 30 14:30:00 crc kubenswrapper[4840]: I0930 14:30:00.332683 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7glv\" (UniqueName: \"kubernetes.io/projected/b27f0395-be2f-431c-baac-5eaeef3ab6c2-kube-api-access-s7glv\") pod \"collect-profiles-29320710-vtk5c\" (UID: \"b27f0395-be2f-431c-baac-5eaeef3ab6c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320710-vtk5c" Sep 30 14:30:00 crc kubenswrapper[4840]: I0930 14:30:00.333270 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b27f0395-be2f-431c-baac-5eaeef3ab6c2-config-volume\") pod \"collect-profiles-29320710-vtk5c\" (UID: \"b27f0395-be2f-431c-baac-5eaeef3ab6c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320710-vtk5c" Sep 30 14:30:00 crc kubenswrapper[4840]: I0930 14:30:00.333930 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b27f0395-be2f-431c-baac-5eaeef3ab6c2-config-volume\") pod \"collect-profiles-29320710-vtk5c\" (UID: \"b27f0395-be2f-431c-baac-5eaeef3ab6c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320710-vtk5c" Sep 30 14:30:00 crc kubenswrapper[4840]: I0930 14:30:00.339361 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b27f0395-be2f-431c-baac-5eaeef3ab6c2-secret-volume\") pod \"collect-profiles-29320710-vtk5c\" (UID: \"b27f0395-be2f-431c-baac-5eaeef3ab6c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320710-vtk5c" Sep 30 14:30:00 crc kubenswrapper[4840]: I0930 14:30:00.364302 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7glv\" (UniqueName: \"kubernetes.io/projected/b27f0395-be2f-431c-baac-5eaeef3ab6c2-kube-api-access-s7glv\") pod \"collect-profiles-29320710-vtk5c\" (UID: \"b27f0395-be2f-431c-baac-5eaeef3ab6c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320710-vtk5c" Sep 30 14:30:00 crc kubenswrapper[4840]: I0930 14:30:00.484162 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320710-vtk5c" Sep 30 14:30:00 crc kubenswrapper[4840]: I0930 14:30:00.914468 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320710-vtk5c"] Sep 30 14:30:00 crc kubenswrapper[4840]: W0930 14:30:00.917215 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb27f0395_be2f_431c_baac_5eaeef3ab6c2.slice/crio-b893b6a8d41420fdaae2d2751ddc841c255265719acb9ea466812089411b4b18 WatchSource:0}: Error finding container b893b6a8d41420fdaae2d2751ddc841c255265719acb9ea466812089411b4b18: Status 404 returned error can't find the container with id b893b6a8d41420fdaae2d2751ddc841c255265719acb9ea466812089411b4b18 Sep 30 14:30:01 crc kubenswrapper[4840]: I0930 14:30:01.750790 4840 generic.go:334] "Generic (PLEG): container finished" podID="b27f0395-be2f-431c-baac-5eaeef3ab6c2" containerID="194fd6ee285c96d7773acd77958b84b8169363971585dd51467e481732ec9457" exitCode=0 Sep 30 14:30:01 crc kubenswrapper[4840]: I0930 14:30:01.750839 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320710-vtk5c" event={"ID":"b27f0395-be2f-431c-baac-5eaeef3ab6c2","Type":"ContainerDied","Data":"194fd6ee285c96d7773acd77958b84b8169363971585dd51467e481732ec9457"} Sep 30 14:30:01 crc kubenswrapper[4840]: I0930 14:30:01.751083 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320710-vtk5c" event={"ID":"b27f0395-be2f-431c-baac-5eaeef3ab6c2","Type":"ContainerStarted","Data":"b893b6a8d41420fdaae2d2751ddc841c255265719acb9ea466812089411b4b18"} Sep 30 14:30:03 crc kubenswrapper[4840]: I0930 14:30:03.094385 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320710-vtk5c" Sep 30 14:30:03 crc kubenswrapper[4840]: I0930 14:30:03.279852 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b27f0395-be2f-431c-baac-5eaeef3ab6c2-secret-volume\") pod \"b27f0395-be2f-431c-baac-5eaeef3ab6c2\" (UID: \"b27f0395-be2f-431c-baac-5eaeef3ab6c2\") " Sep 30 14:30:03 crc kubenswrapper[4840]: I0930 14:30:03.280067 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7glv\" (UniqueName: \"kubernetes.io/projected/b27f0395-be2f-431c-baac-5eaeef3ab6c2-kube-api-access-s7glv\") pod \"b27f0395-be2f-431c-baac-5eaeef3ab6c2\" (UID: \"b27f0395-be2f-431c-baac-5eaeef3ab6c2\") " Sep 30 14:30:03 crc kubenswrapper[4840]: I0930 14:30:03.280310 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b27f0395-be2f-431c-baac-5eaeef3ab6c2-config-volume\") pod \"b27f0395-be2f-431c-baac-5eaeef3ab6c2\" (UID: \"b27f0395-be2f-431c-baac-5eaeef3ab6c2\") " Sep 30 14:30:03 crc kubenswrapper[4840]: I0930 14:30:03.281216 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b27f0395-be2f-431c-baac-5eaeef3ab6c2-config-volume" (OuterVolumeSpecName: "config-volume") pod "b27f0395-be2f-431c-baac-5eaeef3ab6c2" (UID: "b27f0395-be2f-431c-baac-5eaeef3ab6c2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:30:03 crc kubenswrapper[4840]: I0930 14:30:03.286628 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b27f0395-be2f-431c-baac-5eaeef3ab6c2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b27f0395-be2f-431c-baac-5eaeef3ab6c2" (UID: "b27f0395-be2f-431c-baac-5eaeef3ab6c2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:30:03 crc kubenswrapper[4840]: I0930 14:30:03.295238 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b27f0395-be2f-431c-baac-5eaeef3ab6c2-kube-api-access-s7glv" (OuterVolumeSpecName: "kube-api-access-s7glv") pod "b27f0395-be2f-431c-baac-5eaeef3ab6c2" (UID: "b27f0395-be2f-431c-baac-5eaeef3ab6c2"). InnerVolumeSpecName "kube-api-access-s7glv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:30:03 crc kubenswrapper[4840]: I0930 14:30:03.383802 4840 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b27f0395-be2f-431c-baac-5eaeef3ab6c2-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 30 14:30:03 crc kubenswrapper[4840]: I0930 14:30:03.384385 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7glv\" (UniqueName: \"kubernetes.io/projected/b27f0395-be2f-431c-baac-5eaeef3ab6c2-kube-api-access-s7glv\") on node \"crc\" DevicePath \"\"" Sep 30 14:30:03 crc kubenswrapper[4840]: I0930 14:30:03.384404 4840 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b27f0395-be2f-431c-baac-5eaeef3ab6c2-config-volume\") on node \"crc\" DevicePath \"\"" Sep 30 14:30:03 crc kubenswrapper[4840]: I0930 14:30:03.768579 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320710-vtk5c" event={"ID":"b27f0395-be2f-431c-baac-5eaeef3ab6c2","Type":"ContainerDied","Data":"b893b6a8d41420fdaae2d2751ddc841c255265719acb9ea466812089411b4b18"} Sep 30 14:30:03 crc kubenswrapper[4840]: I0930 14:30:03.768613 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b893b6a8d41420fdaae2d2751ddc841c255265719acb9ea466812089411b4b18" Sep 30 14:30:03 crc kubenswrapper[4840]: I0930 14:30:03.769014 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320710-vtk5c" Sep 30 14:30:04 crc kubenswrapper[4840]: I0930 14:30:04.165222 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320665-rkbh5"] Sep 30 14:30:04 crc kubenswrapper[4840]: I0930 14:30:04.172540 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320665-rkbh5"] Sep 30 14:30:06 crc kubenswrapper[4840]: I0930 14:30:06.127592 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="850a2c18-882c-495f-a7e8-743e425a8b61" path="/var/lib/kubelet/pods/850a2c18-882c-495f-a7e8-743e425a8b61/volumes" Sep 30 14:30:15 crc kubenswrapper[4840]: I0930 14:30:15.184624 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7d4fg"] Sep 30 14:30:15 crc kubenswrapper[4840]: E0930 14:30:15.185700 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b27f0395-be2f-431c-baac-5eaeef3ab6c2" containerName="collect-profiles" Sep 30 14:30:15 crc kubenswrapper[4840]: I0930 14:30:15.185718 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="b27f0395-be2f-431c-baac-5eaeef3ab6c2" containerName="collect-profiles" Sep 30 14:30:15 crc kubenswrapper[4840]: I0930 14:30:15.185945 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="b27f0395-be2f-431c-baac-5eaeef3ab6c2" containerName="collect-profiles" Sep 30 14:30:15 crc kubenswrapper[4840]: I0930 14:30:15.187798 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7d4fg" Sep 30 14:30:15 crc kubenswrapper[4840]: I0930 14:30:15.192650 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7d4fg"] Sep 30 14:30:15 crc kubenswrapper[4840]: I0930 14:30:15.302025 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b72e1ec4-e867-45f4-90d3-07a69fa8ee01-utilities\") pod \"certified-operators-7d4fg\" (UID: \"b72e1ec4-e867-45f4-90d3-07a69fa8ee01\") " pod="openshift-marketplace/certified-operators-7d4fg" Sep 30 14:30:15 crc kubenswrapper[4840]: I0930 14:30:15.302078 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fr4zv\" (UniqueName: \"kubernetes.io/projected/b72e1ec4-e867-45f4-90d3-07a69fa8ee01-kube-api-access-fr4zv\") pod \"certified-operators-7d4fg\" (UID: \"b72e1ec4-e867-45f4-90d3-07a69fa8ee01\") " pod="openshift-marketplace/certified-operators-7d4fg" Sep 30 14:30:15 crc kubenswrapper[4840]: I0930 14:30:15.302134 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b72e1ec4-e867-45f4-90d3-07a69fa8ee01-catalog-content\") pod \"certified-operators-7d4fg\" (UID: \"b72e1ec4-e867-45f4-90d3-07a69fa8ee01\") " pod="openshift-marketplace/certified-operators-7d4fg" Sep 30 14:30:15 crc kubenswrapper[4840]: I0930 14:30:15.404114 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b72e1ec4-e867-45f4-90d3-07a69fa8ee01-utilities\") pod \"certified-operators-7d4fg\" (UID: \"b72e1ec4-e867-45f4-90d3-07a69fa8ee01\") " pod="openshift-marketplace/certified-operators-7d4fg" Sep 30 14:30:15 crc kubenswrapper[4840]: I0930 14:30:15.404165 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fr4zv\" (UniqueName: \"kubernetes.io/projected/b72e1ec4-e867-45f4-90d3-07a69fa8ee01-kube-api-access-fr4zv\") pod \"certified-operators-7d4fg\" (UID: \"b72e1ec4-e867-45f4-90d3-07a69fa8ee01\") " pod="openshift-marketplace/certified-operators-7d4fg" Sep 30 14:30:15 crc kubenswrapper[4840]: I0930 14:30:15.404249 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b72e1ec4-e867-45f4-90d3-07a69fa8ee01-catalog-content\") pod \"certified-operators-7d4fg\" (UID: \"b72e1ec4-e867-45f4-90d3-07a69fa8ee01\") " pod="openshift-marketplace/certified-operators-7d4fg" Sep 30 14:30:15 crc kubenswrapper[4840]: I0930 14:30:15.404839 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b72e1ec4-e867-45f4-90d3-07a69fa8ee01-catalog-content\") pod \"certified-operators-7d4fg\" (UID: \"b72e1ec4-e867-45f4-90d3-07a69fa8ee01\") " pod="openshift-marketplace/certified-operators-7d4fg" Sep 30 14:30:15 crc kubenswrapper[4840]: I0930 14:30:15.404855 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b72e1ec4-e867-45f4-90d3-07a69fa8ee01-utilities\") pod \"certified-operators-7d4fg\" (UID: \"b72e1ec4-e867-45f4-90d3-07a69fa8ee01\") " pod="openshift-marketplace/certified-operators-7d4fg" Sep 30 14:30:15 crc kubenswrapper[4840]: I0930 14:30:15.444159 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fr4zv\" (UniqueName: \"kubernetes.io/projected/b72e1ec4-e867-45f4-90d3-07a69fa8ee01-kube-api-access-fr4zv\") pod \"certified-operators-7d4fg\" (UID: \"b72e1ec4-e867-45f4-90d3-07a69fa8ee01\") " pod="openshift-marketplace/certified-operators-7d4fg" Sep 30 14:30:15 crc kubenswrapper[4840]: I0930 14:30:15.519317 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7d4fg" Sep 30 14:30:16 crc kubenswrapper[4840]: I0930 14:30:16.046411 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7d4fg"] Sep 30 14:30:16 crc kubenswrapper[4840]: W0930 14:30:16.049236 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb72e1ec4_e867_45f4_90d3_07a69fa8ee01.slice/crio-19e38a173f5d7aec4c8fa78f4a48e357f15828401fd2e4cdc3fe7ca43ccc6115 WatchSource:0}: Error finding container 19e38a173f5d7aec4c8fa78f4a48e357f15828401fd2e4cdc3fe7ca43ccc6115: Status 404 returned error can't find the container with id 19e38a173f5d7aec4c8fa78f4a48e357f15828401fd2e4cdc3fe7ca43ccc6115 Sep 30 14:30:16 crc kubenswrapper[4840]: I0930 14:30:16.877865 4840 generic.go:334] "Generic (PLEG): container finished" podID="b72e1ec4-e867-45f4-90d3-07a69fa8ee01" containerID="b2ff9b96cffaf519fa30d752f4ddc5af5a13a8d41ef89e6b8323047499bda60d" exitCode=0 Sep 30 14:30:16 crc kubenswrapper[4840]: I0930 14:30:16.877965 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7d4fg" event={"ID":"b72e1ec4-e867-45f4-90d3-07a69fa8ee01","Type":"ContainerDied","Data":"b2ff9b96cffaf519fa30d752f4ddc5af5a13a8d41ef89e6b8323047499bda60d"} Sep 30 14:30:16 crc kubenswrapper[4840]: I0930 14:30:16.878146 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7d4fg" event={"ID":"b72e1ec4-e867-45f4-90d3-07a69fa8ee01","Type":"ContainerStarted","Data":"19e38a173f5d7aec4c8fa78f4a48e357f15828401fd2e4cdc3fe7ca43ccc6115"} Sep 30 14:30:17 crc kubenswrapper[4840]: I0930 14:30:17.889571 4840 generic.go:334] "Generic (PLEG): container finished" podID="b72e1ec4-e867-45f4-90d3-07a69fa8ee01" containerID="c7fdb67fe522e160848655cc5b57c6ef8489745350bcae85f46872a3fe010d57" exitCode=0 Sep 30 14:30:17 crc kubenswrapper[4840]: I0930 14:30:17.889684 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7d4fg" event={"ID":"b72e1ec4-e867-45f4-90d3-07a69fa8ee01","Type":"ContainerDied","Data":"c7fdb67fe522e160848655cc5b57c6ef8489745350bcae85f46872a3fe010d57"} Sep 30 14:30:19 crc kubenswrapper[4840]: I0930 14:30:19.910259 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7d4fg" event={"ID":"b72e1ec4-e867-45f4-90d3-07a69fa8ee01","Type":"ContainerStarted","Data":"a46dcbe25cf212e9f7fbc1ac65aa3b8131c021398147f68a4b264a00dd253e77"} Sep 30 14:30:19 crc kubenswrapper[4840]: I0930 14:30:19.936544 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7d4fg" podStartSLOduration=3.2023227 podStartE2EDuration="4.936522439s" podCreationTimestamp="2025-09-30 14:30:15 +0000 UTC" firstStartedPulling="2025-09-30 14:30:16.879494954 +0000 UTC m=+2045.508581377" lastFinishedPulling="2025-09-30 14:30:18.613694693 +0000 UTC m=+2047.242781116" observedRunningTime="2025-09-30 14:30:19.928194296 +0000 UTC m=+2048.557280729" watchObservedRunningTime="2025-09-30 14:30:19.936522439 +0000 UTC m=+2048.565608862" Sep 30 14:30:25 crc kubenswrapper[4840]: I0930 14:30:25.520073 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7d4fg" Sep 30 14:30:25 crc kubenswrapper[4840]: I0930 14:30:25.520648 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7d4fg" Sep 30 14:30:25 crc kubenswrapper[4840]: I0930 14:30:25.565386 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7d4fg" Sep 30 14:30:26 crc kubenswrapper[4840]: I0930 14:30:26.014682 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7d4fg" Sep 30 14:30:26 crc kubenswrapper[4840]: I0930 14:30:26.065899 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7d4fg"] Sep 30 14:30:27 crc kubenswrapper[4840]: I0930 14:30:27.985960 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7d4fg" podUID="b72e1ec4-e867-45f4-90d3-07a69fa8ee01" containerName="registry-server" containerID="cri-o://a46dcbe25cf212e9f7fbc1ac65aa3b8131c021398147f68a4b264a00dd253e77" gracePeriod=2 Sep 30 14:30:28 crc kubenswrapper[4840]: I0930 14:30:28.424870 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7d4fg" Sep 30 14:30:28 crc kubenswrapper[4840]: I0930 14:30:28.468941 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b72e1ec4-e867-45f4-90d3-07a69fa8ee01-utilities\") pod \"b72e1ec4-e867-45f4-90d3-07a69fa8ee01\" (UID: \"b72e1ec4-e867-45f4-90d3-07a69fa8ee01\") " Sep 30 14:30:28 crc kubenswrapper[4840]: I0930 14:30:28.469079 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fr4zv\" (UniqueName: \"kubernetes.io/projected/b72e1ec4-e867-45f4-90d3-07a69fa8ee01-kube-api-access-fr4zv\") pod \"b72e1ec4-e867-45f4-90d3-07a69fa8ee01\" (UID: \"b72e1ec4-e867-45f4-90d3-07a69fa8ee01\") " Sep 30 14:30:28 crc kubenswrapper[4840]: I0930 14:30:28.469102 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b72e1ec4-e867-45f4-90d3-07a69fa8ee01-catalog-content\") pod \"b72e1ec4-e867-45f4-90d3-07a69fa8ee01\" (UID: \"b72e1ec4-e867-45f4-90d3-07a69fa8ee01\") " Sep 30 14:30:28 crc kubenswrapper[4840]: I0930 14:30:28.469564 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b72e1ec4-e867-45f4-90d3-07a69fa8ee01-utilities" (OuterVolumeSpecName: "utilities") pod "b72e1ec4-e867-45f4-90d3-07a69fa8ee01" (UID: "b72e1ec4-e867-45f4-90d3-07a69fa8ee01"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:30:28 crc kubenswrapper[4840]: I0930 14:30:28.469696 4840 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b72e1ec4-e867-45f4-90d3-07a69fa8ee01-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:30:28 crc kubenswrapper[4840]: I0930 14:30:28.474463 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b72e1ec4-e867-45f4-90d3-07a69fa8ee01-kube-api-access-fr4zv" (OuterVolumeSpecName: "kube-api-access-fr4zv") pod "b72e1ec4-e867-45f4-90d3-07a69fa8ee01" (UID: "b72e1ec4-e867-45f4-90d3-07a69fa8ee01"). InnerVolumeSpecName "kube-api-access-fr4zv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:30:28 crc kubenswrapper[4840]: I0930 14:30:28.506622 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b72e1ec4-e867-45f4-90d3-07a69fa8ee01-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b72e1ec4-e867-45f4-90d3-07a69fa8ee01" (UID: "b72e1ec4-e867-45f4-90d3-07a69fa8ee01"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:30:28 crc kubenswrapper[4840]: I0930 14:30:28.571771 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fr4zv\" (UniqueName: \"kubernetes.io/projected/b72e1ec4-e867-45f4-90d3-07a69fa8ee01-kube-api-access-fr4zv\") on node \"crc\" DevicePath \"\"" Sep 30 14:30:28 crc kubenswrapper[4840]: I0930 14:30:28.571808 4840 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b72e1ec4-e867-45f4-90d3-07a69fa8ee01-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:30:28 crc kubenswrapper[4840]: I0930 14:30:28.999186 4840 generic.go:334] "Generic (PLEG): container finished" podID="b72e1ec4-e867-45f4-90d3-07a69fa8ee01" containerID="a46dcbe25cf212e9f7fbc1ac65aa3b8131c021398147f68a4b264a00dd253e77" exitCode=0 Sep 30 14:30:28 crc kubenswrapper[4840]: I0930 14:30:28.999257 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7d4fg" event={"ID":"b72e1ec4-e867-45f4-90d3-07a69fa8ee01","Type":"ContainerDied","Data":"a46dcbe25cf212e9f7fbc1ac65aa3b8131c021398147f68a4b264a00dd253e77"} Sep 30 14:30:28 crc kubenswrapper[4840]: I0930 14:30:28.999334 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7d4fg" event={"ID":"b72e1ec4-e867-45f4-90d3-07a69fa8ee01","Type":"ContainerDied","Data":"19e38a173f5d7aec4c8fa78f4a48e357f15828401fd2e4cdc3fe7ca43ccc6115"} Sep 30 14:30:28 crc kubenswrapper[4840]: I0930 14:30:28.999362 4840 scope.go:117] "RemoveContainer" containerID="a46dcbe25cf212e9f7fbc1ac65aa3b8131c021398147f68a4b264a00dd253e77" Sep 30 14:30:28 crc kubenswrapper[4840]: I0930 14:30:28.999365 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7d4fg" Sep 30 14:30:29 crc kubenswrapper[4840]: I0930 14:30:29.016764 4840 scope.go:117] "RemoveContainer" containerID="c7fdb67fe522e160848655cc5b57c6ef8489745350bcae85f46872a3fe010d57" Sep 30 14:30:29 crc kubenswrapper[4840]: I0930 14:30:29.047716 4840 scope.go:117] "RemoveContainer" containerID="b2ff9b96cffaf519fa30d752f4ddc5af5a13a8d41ef89e6b8323047499bda60d" Sep 30 14:30:29 crc kubenswrapper[4840]: I0930 14:30:29.089065 4840 scope.go:117] "RemoveContainer" containerID="a46dcbe25cf212e9f7fbc1ac65aa3b8131c021398147f68a4b264a00dd253e77" Sep 30 14:30:29 crc kubenswrapper[4840]: E0930 14:30:29.089736 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a46dcbe25cf212e9f7fbc1ac65aa3b8131c021398147f68a4b264a00dd253e77\": container with ID starting with a46dcbe25cf212e9f7fbc1ac65aa3b8131c021398147f68a4b264a00dd253e77 not found: ID does not exist" containerID="a46dcbe25cf212e9f7fbc1ac65aa3b8131c021398147f68a4b264a00dd253e77" Sep 30 14:30:29 crc kubenswrapper[4840]: I0930 14:30:29.089787 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a46dcbe25cf212e9f7fbc1ac65aa3b8131c021398147f68a4b264a00dd253e77"} err="failed to get container status \"a46dcbe25cf212e9f7fbc1ac65aa3b8131c021398147f68a4b264a00dd253e77\": rpc error: code = NotFound desc = could not find container \"a46dcbe25cf212e9f7fbc1ac65aa3b8131c021398147f68a4b264a00dd253e77\": container with ID starting with a46dcbe25cf212e9f7fbc1ac65aa3b8131c021398147f68a4b264a00dd253e77 not found: ID does not exist" Sep 30 14:30:29 crc kubenswrapper[4840]: I0930 14:30:29.089828 4840 scope.go:117] "RemoveContainer" containerID="c7fdb67fe522e160848655cc5b57c6ef8489745350bcae85f46872a3fe010d57" Sep 30 14:30:29 crc kubenswrapper[4840]: E0930 14:30:29.090338 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7fdb67fe522e160848655cc5b57c6ef8489745350bcae85f46872a3fe010d57\": container with ID starting with c7fdb67fe522e160848655cc5b57c6ef8489745350bcae85f46872a3fe010d57 not found: ID does not exist" containerID="c7fdb67fe522e160848655cc5b57c6ef8489745350bcae85f46872a3fe010d57" Sep 30 14:30:29 crc kubenswrapper[4840]: I0930 14:30:29.090519 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7fdb67fe522e160848655cc5b57c6ef8489745350bcae85f46872a3fe010d57"} err="failed to get container status \"c7fdb67fe522e160848655cc5b57c6ef8489745350bcae85f46872a3fe010d57\": rpc error: code = NotFound desc = could not find container \"c7fdb67fe522e160848655cc5b57c6ef8489745350bcae85f46872a3fe010d57\": container with ID starting with c7fdb67fe522e160848655cc5b57c6ef8489745350bcae85f46872a3fe010d57 not found: ID does not exist" Sep 30 14:30:29 crc kubenswrapper[4840]: I0930 14:30:29.090686 4840 scope.go:117] "RemoveContainer" containerID="b2ff9b96cffaf519fa30d752f4ddc5af5a13a8d41ef89e6b8323047499bda60d" Sep 30 14:30:29 crc kubenswrapper[4840]: E0930 14:30:29.091253 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2ff9b96cffaf519fa30d752f4ddc5af5a13a8d41ef89e6b8323047499bda60d\": container with ID starting with b2ff9b96cffaf519fa30d752f4ddc5af5a13a8d41ef89e6b8323047499bda60d not found: ID does not exist" containerID="b2ff9b96cffaf519fa30d752f4ddc5af5a13a8d41ef89e6b8323047499bda60d" Sep 30 14:30:29 crc kubenswrapper[4840]: I0930 14:30:29.091281 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2ff9b96cffaf519fa30d752f4ddc5af5a13a8d41ef89e6b8323047499bda60d"} err="failed to get container status \"b2ff9b96cffaf519fa30d752f4ddc5af5a13a8d41ef89e6b8323047499bda60d\": rpc error: code = NotFound desc = could not find container \"b2ff9b96cffaf519fa30d752f4ddc5af5a13a8d41ef89e6b8323047499bda60d\": container with ID starting with b2ff9b96cffaf519fa30d752f4ddc5af5a13a8d41ef89e6b8323047499bda60d not found: ID does not exist" Sep 30 14:30:29 crc kubenswrapper[4840]: I0930 14:30:29.097802 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7d4fg"] Sep 30 14:30:29 crc kubenswrapper[4840]: I0930 14:30:29.109784 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7d4fg"] Sep 30 14:30:30 crc kubenswrapper[4840]: I0930 14:30:30.128192 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b72e1ec4-e867-45f4-90d3-07a69fa8ee01" path="/var/lib/kubelet/pods/b72e1ec4-e867-45f4-90d3-07a69fa8ee01/volumes" Sep 30 14:30:35 crc kubenswrapper[4840]: I0930 14:30:35.235300 4840 scope.go:117] "RemoveContainer" containerID="bc5d3b48314ba858d8d83c52b612834e9bda381052fd43502a08ffde2724327d" Sep 30 14:31:51 crc kubenswrapper[4840]: I0930 14:31:51.872423 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:31:51 crc kubenswrapper[4840]: I0930 14:31:51.873688 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:32:21 crc kubenswrapper[4840]: I0930 14:32:21.871686 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:32:21 crc kubenswrapper[4840]: I0930 14:32:21.872156 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:32:51 crc kubenswrapper[4840]: I0930 14:32:51.872448 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:32:51 crc kubenswrapper[4840]: I0930 14:32:51.873291 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:32:51 crc kubenswrapper[4840]: I0930 14:32:51.873443 4840 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-747gk" Sep 30 14:32:51 crc kubenswrapper[4840]: I0930 14:32:51.875787 4840 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6539ed77e1c5a86ff8a57fda399610c4bb26dcd9ac75c7c7b74f7ff71d0bfc80"} pod="openshift-machine-config-operator/machine-config-daemon-747gk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 14:32:51 crc kubenswrapper[4840]: I0930 14:32:51.875868 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" containerID="cri-o://6539ed77e1c5a86ff8a57fda399610c4bb26dcd9ac75c7c7b74f7ff71d0bfc80" gracePeriod=600 Sep 30 14:32:52 crc kubenswrapper[4840]: I0930 14:32:52.237842 4840 generic.go:334] "Generic (PLEG): container finished" podID="10e8b890-7f20-4a36-8e03-898620cf599a" containerID="6539ed77e1c5a86ff8a57fda399610c4bb26dcd9ac75c7c7b74f7ff71d0bfc80" exitCode=0 Sep 30 14:32:52 crc kubenswrapper[4840]: I0930 14:32:52.237932 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" event={"ID":"10e8b890-7f20-4a36-8e03-898620cf599a","Type":"ContainerDied","Data":"6539ed77e1c5a86ff8a57fda399610c4bb26dcd9ac75c7c7b74f7ff71d0bfc80"} Sep 30 14:32:52 crc kubenswrapper[4840]: I0930 14:32:52.238449 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" event={"ID":"10e8b890-7f20-4a36-8e03-898620cf599a","Type":"ContainerStarted","Data":"e1e2bfc7f9537971d768f1b7183216797522532290c5822edafa8e0b5232aa98"} Sep 30 14:32:52 crc kubenswrapper[4840]: I0930 14:32:52.238484 4840 scope.go:117] "RemoveContainer" containerID="74b12cc9318ac99161f1b2541861d89b6100ba9a986e5863f53ae26c2a8953e7" Sep 30 14:33:19 crc kubenswrapper[4840]: I0930 14:33:19.998722 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nxxb8"] Sep 30 14:33:19 crc kubenswrapper[4840]: E0930 14:33:19.999693 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b72e1ec4-e867-45f4-90d3-07a69fa8ee01" containerName="registry-server" Sep 30 14:33:19 crc kubenswrapper[4840]: I0930 14:33:19.999711 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="b72e1ec4-e867-45f4-90d3-07a69fa8ee01" containerName="registry-server" Sep 30 14:33:19 crc kubenswrapper[4840]: E0930 14:33:19.999731 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b72e1ec4-e867-45f4-90d3-07a69fa8ee01" containerName="extract-utilities" Sep 30 14:33:19 crc kubenswrapper[4840]: I0930 14:33:19.999739 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="b72e1ec4-e867-45f4-90d3-07a69fa8ee01" containerName="extract-utilities" Sep 30 14:33:19 crc kubenswrapper[4840]: E0930 14:33:19.999766 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b72e1ec4-e867-45f4-90d3-07a69fa8ee01" containerName="extract-content" Sep 30 14:33:19 crc kubenswrapper[4840]: I0930 14:33:19.999773 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="b72e1ec4-e867-45f4-90d3-07a69fa8ee01" containerName="extract-content" Sep 30 14:33:20 crc kubenswrapper[4840]: I0930 14:33:20.000019 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="b72e1ec4-e867-45f4-90d3-07a69fa8ee01" containerName="registry-server" Sep 30 14:33:20 crc kubenswrapper[4840]: I0930 14:33:20.001676 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nxxb8" Sep 30 14:33:20 crc kubenswrapper[4840]: I0930 14:33:20.014477 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nxxb8"] Sep 30 14:33:20 crc kubenswrapper[4840]: I0930 14:33:20.151454 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98144662-dce4-4cc3-a871-112401e8c5d4-utilities\") pod \"redhat-operators-nxxb8\" (UID: \"98144662-dce4-4cc3-a871-112401e8c5d4\") " pod="openshift-marketplace/redhat-operators-nxxb8" Sep 30 14:33:20 crc kubenswrapper[4840]: I0930 14:33:20.151508 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98144662-dce4-4cc3-a871-112401e8c5d4-catalog-content\") pod \"redhat-operators-nxxb8\" (UID: \"98144662-dce4-4cc3-a871-112401e8c5d4\") " pod="openshift-marketplace/redhat-operators-nxxb8" Sep 30 14:33:20 crc kubenswrapper[4840]: I0930 14:33:20.151599 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87hlq\" (UniqueName: \"kubernetes.io/projected/98144662-dce4-4cc3-a871-112401e8c5d4-kube-api-access-87hlq\") pod \"redhat-operators-nxxb8\" (UID: \"98144662-dce4-4cc3-a871-112401e8c5d4\") " pod="openshift-marketplace/redhat-operators-nxxb8" Sep 30 14:33:20 crc kubenswrapper[4840]: I0930 14:33:20.253186 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98144662-dce4-4cc3-a871-112401e8c5d4-utilities\") pod \"redhat-operators-nxxb8\" (UID: \"98144662-dce4-4cc3-a871-112401e8c5d4\") " pod="openshift-marketplace/redhat-operators-nxxb8" Sep 30 14:33:20 crc kubenswrapper[4840]: I0930 14:33:20.253263 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98144662-dce4-4cc3-a871-112401e8c5d4-catalog-content\") pod \"redhat-operators-nxxb8\" (UID: \"98144662-dce4-4cc3-a871-112401e8c5d4\") " pod="openshift-marketplace/redhat-operators-nxxb8" Sep 30 14:33:20 crc kubenswrapper[4840]: I0930 14:33:20.253409 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87hlq\" (UniqueName: \"kubernetes.io/projected/98144662-dce4-4cc3-a871-112401e8c5d4-kube-api-access-87hlq\") pod \"redhat-operators-nxxb8\" (UID: \"98144662-dce4-4cc3-a871-112401e8c5d4\") " pod="openshift-marketplace/redhat-operators-nxxb8" Sep 30 14:33:20 crc kubenswrapper[4840]: I0930 14:33:20.254004 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98144662-dce4-4cc3-a871-112401e8c5d4-catalog-content\") pod \"redhat-operators-nxxb8\" (UID: \"98144662-dce4-4cc3-a871-112401e8c5d4\") " pod="openshift-marketplace/redhat-operators-nxxb8" Sep 30 14:33:20 crc kubenswrapper[4840]: I0930 14:33:20.254073 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98144662-dce4-4cc3-a871-112401e8c5d4-utilities\") pod \"redhat-operators-nxxb8\" (UID: \"98144662-dce4-4cc3-a871-112401e8c5d4\") " pod="openshift-marketplace/redhat-operators-nxxb8" Sep 30 14:33:20 crc kubenswrapper[4840]: I0930 14:33:20.276901 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87hlq\" (UniqueName: \"kubernetes.io/projected/98144662-dce4-4cc3-a871-112401e8c5d4-kube-api-access-87hlq\") pod \"redhat-operators-nxxb8\" (UID: \"98144662-dce4-4cc3-a871-112401e8c5d4\") " pod="openshift-marketplace/redhat-operators-nxxb8" Sep 30 14:33:20 crc kubenswrapper[4840]: I0930 14:33:20.330225 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nxxb8" Sep 30 14:33:20 crc kubenswrapper[4840]: I0930 14:33:20.795104 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nxxb8"] Sep 30 14:33:21 crc kubenswrapper[4840]: I0930 14:33:21.475936 4840 generic.go:334] "Generic (PLEG): container finished" podID="98144662-dce4-4cc3-a871-112401e8c5d4" containerID="6a9742b271da368452360e11080eb3e5131f4d8887982713d6e946a314864e95" exitCode=0 Sep 30 14:33:21 crc kubenswrapper[4840]: I0930 14:33:21.475980 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nxxb8" event={"ID":"98144662-dce4-4cc3-a871-112401e8c5d4","Type":"ContainerDied","Data":"6a9742b271da368452360e11080eb3e5131f4d8887982713d6e946a314864e95"} Sep 30 14:33:21 crc kubenswrapper[4840]: I0930 14:33:21.476208 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nxxb8" event={"ID":"98144662-dce4-4cc3-a871-112401e8c5d4","Type":"ContainerStarted","Data":"f20be1306362b5f7102fed7cd36280bb2f3f070de15c6f2e834c6262cfa17e02"} Sep 30 14:33:21 crc kubenswrapper[4840]: I0930 14:33:21.477866 4840 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 14:33:23 crc kubenswrapper[4840]: I0930 14:33:23.496365 4840 generic.go:334] "Generic (PLEG): container finished" podID="98144662-dce4-4cc3-a871-112401e8c5d4" containerID="32d61710ca653fea86054c13b6533083e768f629b20acad52b7122681e7bdac6" exitCode=0 Sep 30 14:33:23 crc kubenswrapper[4840]: I0930 14:33:23.496466 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nxxb8" event={"ID":"98144662-dce4-4cc3-a871-112401e8c5d4","Type":"ContainerDied","Data":"32d61710ca653fea86054c13b6533083e768f629b20acad52b7122681e7bdac6"} Sep 30 14:33:25 crc kubenswrapper[4840]: I0930 14:33:25.395093 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wz6kw"] Sep 30 14:33:25 crc kubenswrapper[4840]: I0930 14:33:25.397701 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wz6kw" Sep 30 14:33:25 crc kubenswrapper[4840]: I0930 14:33:25.413860 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wz6kw"] Sep 30 14:33:25 crc kubenswrapper[4840]: I0930 14:33:25.465889 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mfvn\" (UniqueName: \"kubernetes.io/projected/c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed-kube-api-access-9mfvn\") pod \"community-operators-wz6kw\" (UID: \"c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed\") " pod="openshift-marketplace/community-operators-wz6kw" Sep 30 14:33:25 crc kubenswrapper[4840]: I0930 14:33:25.466068 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed-catalog-content\") pod \"community-operators-wz6kw\" (UID: \"c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed\") " pod="openshift-marketplace/community-operators-wz6kw" Sep 30 14:33:25 crc kubenswrapper[4840]: I0930 14:33:25.466221 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed-utilities\") pod \"community-operators-wz6kw\" (UID: \"c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed\") " pod="openshift-marketplace/community-operators-wz6kw" Sep 30 14:33:25 crc kubenswrapper[4840]: I0930 14:33:25.514623 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nxxb8" event={"ID":"98144662-dce4-4cc3-a871-112401e8c5d4","Type":"ContainerStarted","Data":"a61eb453fd0ea2e7941da25f3227dd7b8fa6e2af3fed2066fac1f3758b3c9833"} Sep 30 14:33:25 crc kubenswrapper[4840]: I0930 14:33:25.539873 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nxxb8" podStartSLOduration=3.14325005 podStartE2EDuration="6.539852869s" podCreationTimestamp="2025-09-30 14:33:19 +0000 UTC" firstStartedPulling="2025-09-30 14:33:21.477453065 +0000 UTC m=+2230.106539508" lastFinishedPulling="2025-09-30 14:33:24.874055904 +0000 UTC m=+2233.503142327" observedRunningTime="2025-09-30 14:33:25.533544118 +0000 UTC m=+2234.162630541" watchObservedRunningTime="2025-09-30 14:33:25.539852869 +0000 UTC m=+2234.168939312" Sep 30 14:33:25 crc kubenswrapper[4840]: I0930 14:33:25.567877 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed-catalog-content\") pod \"community-operators-wz6kw\" (UID: \"c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed\") " pod="openshift-marketplace/community-operators-wz6kw" Sep 30 14:33:25 crc kubenswrapper[4840]: I0930 14:33:25.568081 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed-utilities\") pod \"community-operators-wz6kw\" (UID: \"c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed\") " pod="openshift-marketplace/community-operators-wz6kw" Sep 30 14:33:25 crc kubenswrapper[4840]: I0930 14:33:25.568157 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mfvn\" (UniqueName: \"kubernetes.io/projected/c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed-kube-api-access-9mfvn\") pod \"community-operators-wz6kw\" (UID: \"c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed\") " pod="openshift-marketplace/community-operators-wz6kw" Sep 30 14:33:25 crc kubenswrapper[4840]: I0930 14:33:25.568761 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed-utilities\") pod \"community-operators-wz6kw\" (UID: \"c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed\") " pod="openshift-marketplace/community-operators-wz6kw" Sep 30 14:33:25 crc kubenswrapper[4840]: I0930 14:33:25.568870 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed-catalog-content\") pod \"community-operators-wz6kw\" (UID: \"c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed\") " pod="openshift-marketplace/community-operators-wz6kw" Sep 30 14:33:25 crc kubenswrapper[4840]: I0930 14:33:25.593254 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mfvn\" (UniqueName: \"kubernetes.io/projected/c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed-kube-api-access-9mfvn\") pod \"community-operators-wz6kw\" (UID: \"c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed\") " pod="openshift-marketplace/community-operators-wz6kw" Sep 30 14:33:25 crc kubenswrapper[4840]: I0930 14:33:25.717280 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wz6kw" Sep 30 14:33:26 crc kubenswrapper[4840]: I0930 14:33:26.244380 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wz6kw"] Sep 30 14:33:26 crc kubenswrapper[4840]: W0930 14:33:26.270945 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc91fbef9_f7c8_4d2a_a3d0_85c10b10e3ed.slice/crio-ef4e742a364a9caeb706eefa0c1e43b15115ef2d38a9b6648897a8d36e11a9a2 WatchSource:0}: Error finding container ef4e742a364a9caeb706eefa0c1e43b15115ef2d38a9b6648897a8d36e11a9a2: Status 404 returned error can't find the container with id ef4e742a364a9caeb706eefa0c1e43b15115ef2d38a9b6648897a8d36e11a9a2 Sep 30 14:33:26 crc kubenswrapper[4840]: I0930 14:33:26.527796 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wz6kw" event={"ID":"c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed","Type":"ContainerStarted","Data":"8e883ffb46d74fd83a1a4133c9b3e034f45f3a723e99ef9e0bb2fdc5eec1a31f"} Sep 30 14:33:26 crc kubenswrapper[4840]: I0930 14:33:26.527851 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wz6kw" event={"ID":"c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed","Type":"ContainerStarted","Data":"ef4e742a364a9caeb706eefa0c1e43b15115ef2d38a9b6648897a8d36e11a9a2"} Sep 30 14:33:27 crc kubenswrapper[4840]: I0930 14:33:27.538233 4840 generic.go:334] "Generic (PLEG): container finished" podID="c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed" containerID="8e883ffb46d74fd83a1a4133c9b3e034f45f3a723e99ef9e0bb2fdc5eec1a31f" exitCode=0 Sep 30 14:33:27 crc kubenswrapper[4840]: I0930 14:33:27.538353 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wz6kw" event={"ID":"c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed","Type":"ContainerDied","Data":"8e883ffb46d74fd83a1a4133c9b3e034f45f3a723e99ef9e0bb2fdc5eec1a31f"} Sep 30 14:33:29 crc kubenswrapper[4840]: I0930 14:33:29.560155 4840 generic.go:334] "Generic (PLEG): container finished" podID="c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed" containerID="92875ae9b697ff9116634b85c5d4eeaa607e9cc872d9f806d772f48171284e8b" exitCode=0 Sep 30 14:33:29 crc kubenswrapper[4840]: I0930 14:33:29.560273 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wz6kw" event={"ID":"c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed","Type":"ContainerDied","Data":"92875ae9b697ff9116634b85c5d4eeaa607e9cc872d9f806d772f48171284e8b"} Sep 30 14:33:30 crc kubenswrapper[4840]: I0930 14:33:30.331243 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nxxb8" Sep 30 14:33:30 crc kubenswrapper[4840]: I0930 14:33:30.331605 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nxxb8" Sep 30 14:33:30 crc kubenswrapper[4840]: I0930 14:33:30.389894 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nxxb8" Sep 30 14:33:30 crc kubenswrapper[4840]: I0930 14:33:30.588493 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wz6kw" event={"ID":"c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed","Type":"ContainerStarted","Data":"eaeb6dfef832d621c35be0b6fcb4e15358d7f981d06531f643144d4c0ceca37b"} Sep 30 14:33:30 crc kubenswrapper[4840]: I0930 14:33:30.608455 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wz6kw" podStartSLOduration=3.132460046 podStartE2EDuration="5.608439266s" podCreationTimestamp="2025-09-30 14:33:25 +0000 UTC" firstStartedPulling="2025-09-30 14:33:27.540110886 +0000 UTC m=+2236.169197309" lastFinishedPulling="2025-09-30 14:33:30.016090106 +0000 UTC m=+2238.645176529" observedRunningTime="2025-09-30 14:33:30.605565973 +0000 UTC m=+2239.234652406" watchObservedRunningTime="2025-09-30 14:33:30.608439266 +0000 UTC m=+2239.237525689" Sep 30 14:33:30 crc kubenswrapper[4840]: I0930 14:33:30.636411 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nxxb8" Sep 30 14:33:31 crc kubenswrapper[4840]: I0930 14:33:31.186183 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nxxb8"] Sep 30 14:33:32 crc kubenswrapper[4840]: I0930 14:33:32.604436 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nxxb8" podUID="98144662-dce4-4cc3-a871-112401e8c5d4" containerName="registry-server" containerID="cri-o://a61eb453fd0ea2e7941da25f3227dd7b8fa6e2af3fed2066fac1f3758b3c9833" gracePeriod=2 Sep 30 14:33:33 crc kubenswrapper[4840]: I0930 14:33:33.033332 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nxxb8" Sep 30 14:33:33 crc kubenswrapper[4840]: I0930 14:33:33.199665 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87hlq\" (UniqueName: \"kubernetes.io/projected/98144662-dce4-4cc3-a871-112401e8c5d4-kube-api-access-87hlq\") pod \"98144662-dce4-4cc3-a871-112401e8c5d4\" (UID: \"98144662-dce4-4cc3-a871-112401e8c5d4\") " Sep 30 14:33:33 crc kubenswrapper[4840]: I0930 14:33:33.200250 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98144662-dce4-4cc3-a871-112401e8c5d4-catalog-content\") pod \"98144662-dce4-4cc3-a871-112401e8c5d4\" (UID: \"98144662-dce4-4cc3-a871-112401e8c5d4\") " Sep 30 14:33:33 crc kubenswrapper[4840]: I0930 14:33:33.200424 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98144662-dce4-4cc3-a871-112401e8c5d4-utilities\") pod \"98144662-dce4-4cc3-a871-112401e8c5d4\" (UID: \"98144662-dce4-4cc3-a871-112401e8c5d4\") " Sep 30 14:33:33 crc kubenswrapper[4840]: I0930 14:33:33.201383 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98144662-dce4-4cc3-a871-112401e8c5d4-utilities" (OuterVolumeSpecName: "utilities") pod "98144662-dce4-4cc3-a871-112401e8c5d4" (UID: "98144662-dce4-4cc3-a871-112401e8c5d4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:33:33 crc kubenswrapper[4840]: I0930 14:33:33.206533 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98144662-dce4-4cc3-a871-112401e8c5d4-kube-api-access-87hlq" (OuterVolumeSpecName: "kube-api-access-87hlq") pod "98144662-dce4-4cc3-a871-112401e8c5d4" (UID: "98144662-dce4-4cc3-a871-112401e8c5d4"). InnerVolumeSpecName "kube-api-access-87hlq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:33:33 crc kubenswrapper[4840]: I0930 14:33:33.283382 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98144662-dce4-4cc3-a871-112401e8c5d4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "98144662-dce4-4cc3-a871-112401e8c5d4" (UID: "98144662-dce4-4cc3-a871-112401e8c5d4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:33:33 crc kubenswrapper[4840]: I0930 14:33:33.302821 4840 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98144662-dce4-4cc3-a871-112401e8c5d4-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:33:33 crc kubenswrapper[4840]: I0930 14:33:33.302863 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87hlq\" (UniqueName: \"kubernetes.io/projected/98144662-dce4-4cc3-a871-112401e8c5d4-kube-api-access-87hlq\") on node \"crc\" DevicePath \"\"" Sep 30 14:33:33 crc kubenswrapper[4840]: I0930 14:33:33.302881 4840 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98144662-dce4-4cc3-a871-112401e8c5d4-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:33:33 crc kubenswrapper[4840]: I0930 14:33:33.615146 4840 generic.go:334] "Generic (PLEG): container finished" podID="98144662-dce4-4cc3-a871-112401e8c5d4" containerID="a61eb453fd0ea2e7941da25f3227dd7b8fa6e2af3fed2066fac1f3758b3c9833" exitCode=0 Sep 30 14:33:33 crc kubenswrapper[4840]: I0930 14:33:33.615194 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nxxb8" event={"ID":"98144662-dce4-4cc3-a871-112401e8c5d4","Type":"ContainerDied","Data":"a61eb453fd0ea2e7941da25f3227dd7b8fa6e2af3fed2066fac1f3758b3c9833"} Sep 30 14:33:33 crc kubenswrapper[4840]: I0930 14:33:33.615225 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nxxb8" event={"ID":"98144662-dce4-4cc3-a871-112401e8c5d4","Type":"ContainerDied","Data":"f20be1306362b5f7102fed7cd36280bb2f3f070de15c6f2e834c6262cfa17e02"} Sep 30 14:33:33 crc kubenswrapper[4840]: I0930 14:33:33.615245 4840 scope.go:117] "RemoveContainer" containerID="a61eb453fd0ea2e7941da25f3227dd7b8fa6e2af3fed2066fac1f3758b3c9833" Sep 30 14:33:33 crc kubenswrapper[4840]: I0930 14:33:33.615244 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nxxb8" Sep 30 14:33:33 crc kubenswrapper[4840]: I0930 14:33:33.646352 4840 scope.go:117] "RemoveContainer" containerID="32d61710ca653fea86054c13b6533083e768f629b20acad52b7122681e7bdac6" Sep 30 14:33:33 crc kubenswrapper[4840]: I0930 14:33:33.656978 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nxxb8"] Sep 30 14:33:33 crc kubenswrapper[4840]: I0930 14:33:33.664181 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nxxb8"] Sep 30 14:33:33 crc kubenswrapper[4840]: I0930 14:33:33.681333 4840 scope.go:117] "RemoveContainer" containerID="6a9742b271da368452360e11080eb3e5131f4d8887982713d6e946a314864e95" Sep 30 14:33:33 crc kubenswrapper[4840]: I0930 14:33:33.707124 4840 scope.go:117] "RemoveContainer" containerID="a61eb453fd0ea2e7941da25f3227dd7b8fa6e2af3fed2066fac1f3758b3c9833" Sep 30 14:33:33 crc kubenswrapper[4840]: E0930 14:33:33.707606 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a61eb453fd0ea2e7941da25f3227dd7b8fa6e2af3fed2066fac1f3758b3c9833\": container with ID starting with a61eb453fd0ea2e7941da25f3227dd7b8fa6e2af3fed2066fac1f3758b3c9833 not found: ID does not exist" containerID="a61eb453fd0ea2e7941da25f3227dd7b8fa6e2af3fed2066fac1f3758b3c9833" Sep 30 14:33:33 crc kubenswrapper[4840]: I0930 14:33:33.707635 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a61eb453fd0ea2e7941da25f3227dd7b8fa6e2af3fed2066fac1f3758b3c9833"} err="failed to get container status \"a61eb453fd0ea2e7941da25f3227dd7b8fa6e2af3fed2066fac1f3758b3c9833\": rpc error: code = NotFound desc = could not find container \"a61eb453fd0ea2e7941da25f3227dd7b8fa6e2af3fed2066fac1f3758b3c9833\": container with ID starting with a61eb453fd0ea2e7941da25f3227dd7b8fa6e2af3fed2066fac1f3758b3c9833 not found: ID does not exist" Sep 30 14:33:33 crc kubenswrapper[4840]: I0930 14:33:33.707655 4840 scope.go:117] "RemoveContainer" containerID="32d61710ca653fea86054c13b6533083e768f629b20acad52b7122681e7bdac6" Sep 30 14:33:33 crc kubenswrapper[4840]: E0930 14:33:33.708022 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32d61710ca653fea86054c13b6533083e768f629b20acad52b7122681e7bdac6\": container with ID starting with 32d61710ca653fea86054c13b6533083e768f629b20acad52b7122681e7bdac6 not found: ID does not exist" containerID="32d61710ca653fea86054c13b6533083e768f629b20acad52b7122681e7bdac6" Sep 30 14:33:33 crc kubenswrapper[4840]: I0930 14:33:33.708063 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32d61710ca653fea86054c13b6533083e768f629b20acad52b7122681e7bdac6"} err="failed to get container status \"32d61710ca653fea86054c13b6533083e768f629b20acad52b7122681e7bdac6\": rpc error: code = NotFound desc = could not find container \"32d61710ca653fea86054c13b6533083e768f629b20acad52b7122681e7bdac6\": container with ID starting with 32d61710ca653fea86054c13b6533083e768f629b20acad52b7122681e7bdac6 not found: ID does not exist" Sep 30 14:33:33 crc kubenswrapper[4840]: I0930 14:33:33.708090 4840 scope.go:117] "RemoveContainer" containerID="6a9742b271da368452360e11080eb3e5131f4d8887982713d6e946a314864e95" Sep 30 14:33:33 crc kubenswrapper[4840]: E0930 14:33:33.708748 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a9742b271da368452360e11080eb3e5131f4d8887982713d6e946a314864e95\": container with ID starting with 6a9742b271da368452360e11080eb3e5131f4d8887982713d6e946a314864e95 not found: ID does not exist" containerID="6a9742b271da368452360e11080eb3e5131f4d8887982713d6e946a314864e95" Sep 30 14:33:33 crc kubenswrapper[4840]: I0930 14:33:33.708806 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a9742b271da368452360e11080eb3e5131f4d8887982713d6e946a314864e95"} err="failed to get container status \"6a9742b271da368452360e11080eb3e5131f4d8887982713d6e946a314864e95\": rpc error: code = NotFound desc = could not find container \"6a9742b271da368452360e11080eb3e5131f4d8887982713d6e946a314864e95\": container with ID starting with 6a9742b271da368452360e11080eb3e5131f4d8887982713d6e946a314864e95 not found: ID does not exist" Sep 30 14:33:34 crc kubenswrapper[4840]: I0930 14:33:34.129491 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98144662-dce4-4cc3-a871-112401e8c5d4" path="/var/lib/kubelet/pods/98144662-dce4-4cc3-a871-112401e8c5d4/volumes" Sep 30 14:33:35 crc kubenswrapper[4840]: I0930 14:33:35.718391 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wz6kw" Sep 30 14:33:35 crc kubenswrapper[4840]: I0930 14:33:35.719268 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wz6kw" Sep 30 14:33:35 crc kubenswrapper[4840]: I0930 14:33:35.763304 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wz6kw" Sep 30 14:33:36 crc kubenswrapper[4840]: I0930 14:33:36.708351 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wz6kw" Sep 30 14:33:37 crc kubenswrapper[4840]: I0930 14:33:37.591124 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wz6kw"] Sep 30 14:33:38 crc kubenswrapper[4840]: I0930 14:33:38.670385 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wz6kw" podUID="c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed" containerName="registry-server" containerID="cri-o://eaeb6dfef832d621c35be0b6fcb4e15358d7f981d06531f643144d4c0ceca37b" gracePeriod=2 Sep 30 14:33:39 crc kubenswrapper[4840]: I0930 14:33:39.654730 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wz6kw" Sep 30 14:33:39 crc kubenswrapper[4840]: I0930 14:33:39.680517 4840 generic.go:334] "Generic (PLEG): container finished" podID="c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed" containerID="eaeb6dfef832d621c35be0b6fcb4e15358d7f981d06531f643144d4c0ceca37b" exitCode=0 Sep 30 14:33:39 crc kubenswrapper[4840]: I0930 14:33:39.680574 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wz6kw" event={"ID":"c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed","Type":"ContainerDied","Data":"eaeb6dfef832d621c35be0b6fcb4e15358d7f981d06531f643144d4c0ceca37b"} Sep 30 14:33:39 crc kubenswrapper[4840]: I0930 14:33:39.680618 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wz6kw" event={"ID":"c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed","Type":"ContainerDied","Data":"ef4e742a364a9caeb706eefa0c1e43b15115ef2d38a9b6648897a8d36e11a9a2"} Sep 30 14:33:39 crc kubenswrapper[4840]: I0930 14:33:39.680640 4840 scope.go:117] "RemoveContainer" containerID="eaeb6dfef832d621c35be0b6fcb4e15358d7f981d06531f643144d4c0ceca37b" Sep 30 14:33:39 crc kubenswrapper[4840]: I0930 14:33:39.680644 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wz6kw" Sep 30 14:33:39 crc kubenswrapper[4840]: I0930 14:33:39.700189 4840 scope.go:117] "RemoveContainer" containerID="92875ae9b697ff9116634b85c5d4eeaa607e9cc872d9f806d772f48171284e8b" Sep 30 14:33:39 crc kubenswrapper[4840]: I0930 14:33:39.724884 4840 scope.go:117] "RemoveContainer" containerID="8e883ffb46d74fd83a1a4133c9b3e034f45f3a723e99ef9e0bb2fdc5eec1a31f" Sep 30 14:33:39 crc kubenswrapper[4840]: I0930 14:33:39.765808 4840 scope.go:117] "RemoveContainer" containerID="eaeb6dfef832d621c35be0b6fcb4e15358d7f981d06531f643144d4c0ceca37b" Sep 30 14:33:39 crc kubenswrapper[4840]: E0930 14:33:39.766336 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eaeb6dfef832d621c35be0b6fcb4e15358d7f981d06531f643144d4c0ceca37b\": container with ID starting with eaeb6dfef832d621c35be0b6fcb4e15358d7f981d06531f643144d4c0ceca37b not found: ID does not exist" containerID="eaeb6dfef832d621c35be0b6fcb4e15358d7f981d06531f643144d4c0ceca37b" Sep 30 14:33:39 crc kubenswrapper[4840]: I0930 14:33:39.766397 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eaeb6dfef832d621c35be0b6fcb4e15358d7f981d06531f643144d4c0ceca37b"} err="failed to get container status \"eaeb6dfef832d621c35be0b6fcb4e15358d7f981d06531f643144d4c0ceca37b\": rpc error: code = NotFound desc = could not find container \"eaeb6dfef832d621c35be0b6fcb4e15358d7f981d06531f643144d4c0ceca37b\": container with ID starting with eaeb6dfef832d621c35be0b6fcb4e15358d7f981d06531f643144d4c0ceca37b not found: ID does not exist" Sep 30 14:33:39 crc kubenswrapper[4840]: I0930 14:33:39.766424 4840 scope.go:117] "RemoveContainer" containerID="92875ae9b697ff9116634b85c5d4eeaa607e9cc872d9f806d772f48171284e8b" Sep 30 14:33:39 crc kubenswrapper[4840]: E0930 14:33:39.766929 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92875ae9b697ff9116634b85c5d4eeaa607e9cc872d9f806d772f48171284e8b\": container with ID starting with 92875ae9b697ff9116634b85c5d4eeaa607e9cc872d9f806d772f48171284e8b not found: ID does not exist" containerID="92875ae9b697ff9116634b85c5d4eeaa607e9cc872d9f806d772f48171284e8b" Sep 30 14:33:39 crc kubenswrapper[4840]: I0930 14:33:39.766976 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92875ae9b697ff9116634b85c5d4eeaa607e9cc872d9f806d772f48171284e8b"} err="failed to get container status \"92875ae9b697ff9116634b85c5d4eeaa607e9cc872d9f806d772f48171284e8b\": rpc error: code = NotFound desc = could not find container \"92875ae9b697ff9116634b85c5d4eeaa607e9cc872d9f806d772f48171284e8b\": container with ID starting with 92875ae9b697ff9116634b85c5d4eeaa607e9cc872d9f806d772f48171284e8b not found: ID does not exist" Sep 30 14:33:39 crc kubenswrapper[4840]: I0930 14:33:39.767006 4840 scope.go:117] "RemoveContainer" containerID="8e883ffb46d74fd83a1a4133c9b3e034f45f3a723e99ef9e0bb2fdc5eec1a31f" Sep 30 14:33:39 crc kubenswrapper[4840]: E0930 14:33:39.767387 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e883ffb46d74fd83a1a4133c9b3e034f45f3a723e99ef9e0bb2fdc5eec1a31f\": container with ID starting with 8e883ffb46d74fd83a1a4133c9b3e034f45f3a723e99ef9e0bb2fdc5eec1a31f not found: ID does not exist" containerID="8e883ffb46d74fd83a1a4133c9b3e034f45f3a723e99ef9e0bb2fdc5eec1a31f" Sep 30 14:33:39 crc kubenswrapper[4840]: I0930 14:33:39.767421 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e883ffb46d74fd83a1a4133c9b3e034f45f3a723e99ef9e0bb2fdc5eec1a31f"} err="failed to get container status \"8e883ffb46d74fd83a1a4133c9b3e034f45f3a723e99ef9e0bb2fdc5eec1a31f\": rpc error: code = NotFound desc = could not find container \"8e883ffb46d74fd83a1a4133c9b3e034f45f3a723e99ef9e0bb2fdc5eec1a31f\": container with ID starting with 8e883ffb46d74fd83a1a4133c9b3e034f45f3a723e99ef9e0bb2fdc5eec1a31f not found: ID does not exist" Sep 30 14:33:39 crc kubenswrapper[4840]: I0930 14:33:39.823102 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed-catalog-content\") pod \"c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed\" (UID: \"c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed\") " Sep 30 14:33:39 crc kubenswrapper[4840]: I0930 14:33:39.823185 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9mfvn\" (UniqueName: \"kubernetes.io/projected/c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed-kube-api-access-9mfvn\") pod \"c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed\" (UID: \"c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed\") " Sep 30 14:33:39 crc kubenswrapper[4840]: I0930 14:33:39.823367 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed-utilities\") pod \"c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed\" (UID: \"c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed\") " Sep 30 14:33:39 crc kubenswrapper[4840]: I0930 14:33:39.824518 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed-utilities" (OuterVolumeSpecName: "utilities") pod "c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed" (UID: "c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:33:39 crc kubenswrapper[4840]: I0930 14:33:39.828441 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed-kube-api-access-9mfvn" (OuterVolumeSpecName: "kube-api-access-9mfvn") pod "c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed" (UID: "c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed"). InnerVolumeSpecName "kube-api-access-9mfvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:33:39 crc kubenswrapper[4840]: I0930 14:33:39.870468 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed" (UID: "c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:33:39 crc kubenswrapper[4840]: I0930 14:33:39.925181 4840 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:33:39 crc kubenswrapper[4840]: I0930 14:33:39.925220 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9mfvn\" (UniqueName: \"kubernetes.io/projected/c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed-kube-api-access-9mfvn\") on node \"crc\" DevicePath \"\"" Sep 30 14:33:39 crc kubenswrapper[4840]: I0930 14:33:39.925231 4840 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:33:40 crc kubenswrapper[4840]: I0930 14:33:40.009181 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wz6kw"] Sep 30 14:33:40 crc kubenswrapper[4840]: I0930 14:33:40.017145 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wz6kw"] Sep 30 14:33:40 crc kubenswrapper[4840]: I0930 14:33:40.132897 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed" path="/var/lib/kubelet/pods/c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed/volumes" Sep 30 14:33:49 crc kubenswrapper[4840]: I0930 14:33:49.768814 4840 generic.go:334] "Generic (PLEG): container finished" podID="9f66833a-f0de-410b-9459-80203d4291a6" containerID="64a9bd6bceffc16ee4b6db867c83bf1cb58cac34b2aab3d521ee942e2334fd1a" exitCode=0 Sep 30 14:33:49 crc kubenswrapper[4840]: I0930 14:33:49.768861 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m584b" event={"ID":"9f66833a-f0de-410b-9459-80203d4291a6","Type":"ContainerDied","Data":"64a9bd6bceffc16ee4b6db867c83bf1cb58cac34b2aab3d521ee942e2334fd1a"} Sep 30 14:33:51 crc kubenswrapper[4840]: I0930 14:33:51.149347 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m584b" Sep 30 14:33:51 crc kubenswrapper[4840]: I0930 14:33:51.326484 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f66833a-f0de-410b-9459-80203d4291a6-libvirt-combined-ca-bundle\") pod \"9f66833a-f0de-410b-9459-80203d4291a6\" (UID: \"9f66833a-f0de-410b-9459-80203d4291a6\") " Sep 30 14:33:51 crc kubenswrapper[4840]: I0930 14:33:51.326631 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f66833a-f0de-410b-9459-80203d4291a6-inventory\") pod \"9f66833a-f0de-410b-9459-80203d4291a6\" (UID: \"9f66833a-f0de-410b-9459-80203d4291a6\") " Sep 30 14:33:51 crc kubenswrapper[4840]: I0930 14:33:51.326743 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/9f66833a-f0de-410b-9459-80203d4291a6-libvirt-secret-0\") pod \"9f66833a-f0de-410b-9459-80203d4291a6\" (UID: \"9f66833a-f0de-410b-9459-80203d4291a6\") " Sep 30 14:33:51 crc kubenswrapper[4840]: I0930 14:33:51.326872 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f66833a-f0de-410b-9459-80203d4291a6-ssh-key\") pod \"9f66833a-f0de-410b-9459-80203d4291a6\" (UID: \"9f66833a-f0de-410b-9459-80203d4291a6\") " Sep 30 14:33:51 crc kubenswrapper[4840]: I0930 14:33:51.326945 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89j6f\" (UniqueName: \"kubernetes.io/projected/9f66833a-f0de-410b-9459-80203d4291a6-kube-api-access-89j6f\") pod \"9f66833a-f0de-410b-9459-80203d4291a6\" (UID: \"9f66833a-f0de-410b-9459-80203d4291a6\") " Sep 30 14:33:51 crc kubenswrapper[4840]: I0930 14:33:51.334658 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f66833a-f0de-410b-9459-80203d4291a6-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "9f66833a-f0de-410b-9459-80203d4291a6" (UID: "9f66833a-f0de-410b-9459-80203d4291a6"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:33:51 crc kubenswrapper[4840]: I0930 14:33:51.334979 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f66833a-f0de-410b-9459-80203d4291a6-kube-api-access-89j6f" (OuterVolumeSpecName: "kube-api-access-89j6f") pod "9f66833a-f0de-410b-9459-80203d4291a6" (UID: "9f66833a-f0de-410b-9459-80203d4291a6"). InnerVolumeSpecName "kube-api-access-89j6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:33:51 crc kubenswrapper[4840]: I0930 14:33:51.357075 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f66833a-f0de-410b-9459-80203d4291a6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9f66833a-f0de-410b-9459-80203d4291a6" (UID: "9f66833a-f0de-410b-9459-80203d4291a6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:33:51 crc kubenswrapper[4840]: I0930 14:33:51.361065 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f66833a-f0de-410b-9459-80203d4291a6-inventory" (OuterVolumeSpecName: "inventory") pod "9f66833a-f0de-410b-9459-80203d4291a6" (UID: "9f66833a-f0de-410b-9459-80203d4291a6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:33:51 crc kubenswrapper[4840]: I0930 14:33:51.362421 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f66833a-f0de-410b-9459-80203d4291a6-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "9f66833a-f0de-410b-9459-80203d4291a6" (UID: "9f66833a-f0de-410b-9459-80203d4291a6"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:33:51 crc kubenswrapper[4840]: I0930 14:33:51.429305 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89j6f\" (UniqueName: \"kubernetes.io/projected/9f66833a-f0de-410b-9459-80203d4291a6-kube-api-access-89j6f\") on node \"crc\" DevicePath \"\"" Sep 30 14:33:51 crc kubenswrapper[4840]: I0930 14:33:51.429348 4840 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f66833a-f0de-410b-9459-80203d4291a6-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:33:51 crc kubenswrapper[4840]: I0930 14:33:51.429363 4840 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f66833a-f0de-410b-9459-80203d4291a6-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 14:33:51 crc kubenswrapper[4840]: I0930 14:33:51.429377 4840 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/9f66833a-f0de-410b-9459-80203d4291a6-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:33:51 crc kubenswrapper[4840]: I0930 14:33:51.429389 4840 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f66833a-f0de-410b-9459-80203d4291a6-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:33:51 crc kubenswrapper[4840]: I0930 14:33:51.788812 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m584b" event={"ID":"9f66833a-f0de-410b-9459-80203d4291a6","Type":"ContainerDied","Data":"701804470c4d4cca4a5a0da1a8b9c7a3808f8a37e56140dc1807f7029d12bbb2"} Sep 30 14:33:51 crc kubenswrapper[4840]: I0930 14:33:51.788853 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="701804470c4d4cca4a5a0da1a8b9c7a3808f8a37e56140dc1807f7029d12bbb2" Sep 30 14:33:51 crc kubenswrapper[4840]: I0930 14:33:51.788935 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m584b" Sep 30 14:33:51 crc kubenswrapper[4840]: I0930 14:33:51.866726 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-5q27c"] Sep 30 14:33:51 crc kubenswrapper[4840]: E0930 14:33:51.867189 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed" containerName="extract-content" Sep 30 14:33:51 crc kubenswrapper[4840]: I0930 14:33:51.867217 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed" containerName="extract-content" Sep 30 14:33:51 crc kubenswrapper[4840]: E0930 14:33:51.867238 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed" containerName="registry-server" Sep 30 14:33:51 crc kubenswrapper[4840]: I0930 14:33:51.867246 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed" containerName="registry-server" Sep 30 14:33:51 crc kubenswrapper[4840]: E0930 14:33:51.867283 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98144662-dce4-4cc3-a871-112401e8c5d4" containerName="registry-server" Sep 30 14:33:51 crc kubenswrapper[4840]: I0930 14:33:51.867292 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="98144662-dce4-4cc3-a871-112401e8c5d4" containerName="registry-server" Sep 30 14:33:51 crc kubenswrapper[4840]: E0930 14:33:51.867312 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98144662-dce4-4cc3-a871-112401e8c5d4" containerName="extract-content" Sep 30 14:33:51 crc kubenswrapper[4840]: I0930 14:33:51.867322 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="98144662-dce4-4cc3-a871-112401e8c5d4" containerName="extract-content" Sep 30 14:33:51 crc kubenswrapper[4840]: E0930 14:33:51.867347 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed" containerName="extract-utilities" Sep 30 14:33:51 crc kubenswrapper[4840]: I0930 14:33:51.867355 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed" containerName="extract-utilities" Sep 30 14:33:51 crc kubenswrapper[4840]: E0930 14:33:51.867367 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f66833a-f0de-410b-9459-80203d4291a6" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Sep 30 14:33:51 crc kubenswrapper[4840]: I0930 14:33:51.867378 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f66833a-f0de-410b-9459-80203d4291a6" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Sep 30 14:33:51 crc kubenswrapper[4840]: E0930 14:33:51.867394 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98144662-dce4-4cc3-a871-112401e8c5d4" containerName="extract-utilities" Sep 30 14:33:51 crc kubenswrapper[4840]: I0930 14:33:51.867415 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="98144662-dce4-4cc3-a871-112401e8c5d4" containerName="extract-utilities" Sep 30 14:33:51 crc kubenswrapper[4840]: I0930 14:33:51.867664 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="c91fbef9-f7c8-4d2a-a3d0-85c10b10e3ed" containerName="registry-server" Sep 30 14:33:51 crc kubenswrapper[4840]: I0930 14:33:51.867690 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f66833a-f0de-410b-9459-80203d4291a6" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Sep 30 14:33:51 crc kubenswrapper[4840]: I0930 14:33:51.867698 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="98144662-dce4-4cc3-a871-112401e8c5d4" containerName="registry-server" Sep 30 14:33:51 crc kubenswrapper[4840]: I0930 14:33:51.868354 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5q27c" Sep 30 14:33:51 crc kubenswrapper[4840]: I0930 14:33:51.870496 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9zm9z" Sep 30 14:33:51 crc kubenswrapper[4840]: I0930 14:33:51.871150 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 14:33:51 crc kubenswrapper[4840]: I0930 14:33:51.871209 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 14:33:51 crc kubenswrapper[4840]: I0930 14:33:51.871264 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Sep 30 14:33:51 crc kubenswrapper[4840]: I0930 14:33:51.871166 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 14:33:51 crc kubenswrapper[4840]: I0930 14:33:51.871482 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Sep 30 14:33:51 crc kubenswrapper[4840]: I0930 14:33:51.875098 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Sep 30 14:33:51 crc kubenswrapper[4840]: I0930 14:33:51.883117 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-5q27c"] Sep 30 14:33:52 crc kubenswrapper[4840]: I0930 14:33:52.039985 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrmc9\" (UniqueName: \"kubernetes.io/projected/20b6236d-9397-4095-a35a-6c24222b852f-kube-api-access-qrmc9\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5q27c\" (UID: \"20b6236d-9397-4095-a35a-6c24222b852f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5q27c" Sep 30 14:33:52 crc kubenswrapper[4840]: I0930 14:33:52.040077 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5q27c\" (UID: \"20b6236d-9397-4095-a35a-6c24222b852f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5q27c" Sep 30 14:33:52 crc kubenswrapper[4840]: I0930 14:33:52.040156 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5q27c\" (UID: \"20b6236d-9397-4095-a35a-6c24222b852f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5q27c" Sep 30 14:33:52 crc kubenswrapper[4840]: I0930 14:33:52.040207 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5q27c\" (UID: \"20b6236d-9397-4095-a35a-6c24222b852f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5q27c" Sep 30 14:33:52 crc kubenswrapper[4840]: I0930 14:33:52.040242 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5q27c\" (UID: \"20b6236d-9397-4095-a35a-6c24222b852f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5q27c" Sep 30 14:33:52 crc kubenswrapper[4840]: I0930 14:33:52.040327 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5q27c\" (UID: \"20b6236d-9397-4095-a35a-6c24222b852f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5q27c" Sep 30 14:33:52 crc kubenswrapper[4840]: I0930 14:33:52.040355 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5q27c\" (UID: \"20b6236d-9397-4095-a35a-6c24222b852f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5q27c" Sep 30 14:33:52 crc kubenswrapper[4840]: I0930 14:33:52.040400 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5q27c\" (UID: \"20b6236d-9397-4095-a35a-6c24222b852f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5q27c" Sep 30 14:33:52 crc kubenswrapper[4840]: I0930 14:33:52.040427 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/20b6236d-9397-4095-a35a-6c24222b852f-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5q27c\" (UID: \"20b6236d-9397-4095-a35a-6c24222b852f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5q27c" Sep 30 14:33:52 crc kubenswrapper[4840]: I0930 14:33:52.142698 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5q27c\" (UID: \"20b6236d-9397-4095-a35a-6c24222b852f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5q27c" Sep 30 14:33:52 crc kubenswrapper[4840]: I0930 14:33:52.142791 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5q27c\" (UID: \"20b6236d-9397-4095-a35a-6c24222b852f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5q27c" Sep 30 14:33:52 crc kubenswrapper[4840]: I0930 14:33:52.142846 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5q27c\" (UID: \"20b6236d-9397-4095-a35a-6c24222b852f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5q27c" Sep 30 14:33:52 crc kubenswrapper[4840]: I0930 14:33:52.142877 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5q27c\" (UID: \"20b6236d-9397-4095-a35a-6c24222b852f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5q27c" Sep 30 14:33:52 crc kubenswrapper[4840]: I0930 14:33:52.143096 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5q27c\" (UID: \"20b6236d-9397-4095-a35a-6c24222b852f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5q27c" Sep 30 14:33:52 crc kubenswrapper[4840]: I0930 14:33:52.143131 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5q27c\" (UID: \"20b6236d-9397-4095-a35a-6c24222b852f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5q27c" Sep 30 14:33:52 crc kubenswrapper[4840]: I0930 14:33:52.143183 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5q27c\" (UID: \"20b6236d-9397-4095-a35a-6c24222b852f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5q27c" Sep 30 14:33:52 crc kubenswrapper[4840]: I0930 14:33:52.143208 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/20b6236d-9397-4095-a35a-6c24222b852f-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5q27c\" (UID: \"20b6236d-9397-4095-a35a-6c24222b852f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5q27c" Sep 30 14:33:52 crc kubenswrapper[4840]: I0930 14:33:52.143392 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrmc9\" (UniqueName: \"kubernetes.io/projected/20b6236d-9397-4095-a35a-6c24222b852f-kube-api-access-qrmc9\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5q27c\" (UID: \"20b6236d-9397-4095-a35a-6c24222b852f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5q27c" Sep 30 14:33:52 crc kubenswrapper[4840]: I0930 14:33:52.144659 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/20b6236d-9397-4095-a35a-6c24222b852f-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5q27c\" (UID: \"20b6236d-9397-4095-a35a-6c24222b852f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5q27c" Sep 30 14:33:52 crc kubenswrapper[4840]: I0930 14:33:52.147715 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5q27c\" (UID: \"20b6236d-9397-4095-a35a-6c24222b852f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5q27c" Sep 30 14:33:52 crc kubenswrapper[4840]: I0930 14:33:52.148008 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5q27c\" (UID: \"20b6236d-9397-4095-a35a-6c24222b852f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5q27c" Sep 30 14:33:52 crc kubenswrapper[4840]: I0930 14:33:52.148091 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5q27c\" (UID: \"20b6236d-9397-4095-a35a-6c24222b852f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5q27c" Sep 30 14:33:52 crc kubenswrapper[4840]: I0930 14:33:52.148643 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5q27c\" (UID: \"20b6236d-9397-4095-a35a-6c24222b852f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5q27c" Sep 30 14:33:52 crc kubenswrapper[4840]: I0930 14:33:52.149271 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5q27c\" (UID: \"20b6236d-9397-4095-a35a-6c24222b852f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5q27c" Sep 30 14:33:52 crc kubenswrapper[4840]: I0930 14:33:52.149284 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5q27c\" (UID: \"20b6236d-9397-4095-a35a-6c24222b852f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5q27c" Sep 30 14:33:52 crc kubenswrapper[4840]: I0930 14:33:52.150111 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5q27c\" (UID: \"20b6236d-9397-4095-a35a-6c24222b852f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5q27c" Sep 30 14:33:52 crc kubenswrapper[4840]: I0930 14:33:52.163269 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrmc9\" (UniqueName: \"kubernetes.io/projected/20b6236d-9397-4095-a35a-6c24222b852f-kube-api-access-qrmc9\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5q27c\" (UID: \"20b6236d-9397-4095-a35a-6c24222b852f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5q27c" Sep 30 14:33:52 crc kubenswrapper[4840]: I0930 14:33:52.190531 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5q27c" Sep 30 14:33:52 crc kubenswrapper[4840]: I0930 14:33:52.724071 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-5q27c"] Sep 30 14:33:52 crc kubenswrapper[4840]: I0930 14:33:52.799506 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5q27c" event={"ID":"20b6236d-9397-4095-a35a-6c24222b852f","Type":"ContainerStarted","Data":"4dc73b0b5dae9d180187b29a11f347b8ca9690183b66cb6f3e46008d2815475c"} Sep 30 14:33:53 crc kubenswrapper[4840]: I0930 14:33:53.809125 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5q27c" event={"ID":"20b6236d-9397-4095-a35a-6c24222b852f","Type":"ContainerStarted","Data":"3abfd0972e294f9859f00dc0f8bf56da70b63dd26a76e86e1793f33b0d90451a"} Sep 30 14:33:53 crc kubenswrapper[4840]: I0930 14:33:53.830831 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5q27c" podStartSLOduration=2.326760511 podStartE2EDuration="2.830809488s" podCreationTimestamp="2025-09-30 14:33:51 +0000 UTC" firstStartedPulling="2025-09-30 14:33:52.727753891 +0000 UTC m=+2261.356840314" lastFinishedPulling="2025-09-30 14:33:53.231802868 +0000 UTC m=+2261.860889291" observedRunningTime="2025-09-30 14:33:53.823967433 +0000 UTC m=+2262.453053866" watchObservedRunningTime="2025-09-30 14:33:53.830809488 +0000 UTC m=+2262.459895931" Sep 30 14:35:21 crc kubenswrapper[4840]: I0930 14:35:21.871753 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:35:21 crc kubenswrapper[4840]: I0930 14:35:21.872631 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:35:51 crc kubenswrapper[4840]: I0930 14:35:51.871781 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:35:51 crc kubenswrapper[4840]: I0930 14:35:51.872759 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:36:21 crc kubenswrapper[4840]: I0930 14:36:21.871594 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:36:21 crc kubenswrapper[4840]: I0930 14:36:21.872205 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:36:21 crc kubenswrapper[4840]: I0930 14:36:21.872275 4840 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-747gk" Sep 30 14:36:22 crc kubenswrapper[4840]: I0930 14:36:22.051014 4840 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e1e2bfc7f9537971d768f1b7183216797522532290c5822edafa8e0b5232aa98"} pod="openshift-machine-config-operator/machine-config-daemon-747gk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 14:36:22 crc kubenswrapper[4840]: I0930 14:36:22.051159 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" containerID="cri-o://e1e2bfc7f9537971d768f1b7183216797522532290c5822edafa8e0b5232aa98" gracePeriod=600 Sep 30 14:36:22 crc kubenswrapper[4840]: E0930 14:36:22.167686 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:36:23 crc kubenswrapper[4840]: I0930 14:36:23.061338 4840 generic.go:334] "Generic (PLEG): container finished" podID="10e8b890-7f20-4a36-8e03-898620cf599a" containerID="e1e2bfc7f9537971d768f1b7183216797522532290c5822edafa8e0b5232aa98" exitCode=0 Sep 30 14:36:23 crc kubenswrapper[4840]: I0930 14:36:23.061427 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" event={"ID":"10e8b890-7f20-4a36-8e03-898620cf599a","Type":"ContainerDied","Data":"e1e2bfc7f9537971d768f1b7183216797522532290c5822edafa8e0b5232aa98"} Sep 30 14:36:23 crc kubenswrapper[4840]: I0930 14:36:23.061672 4840 scope.go:117] "RemoveContainer" containerID="6539ed77e1c5a86ff8a57fda399610c4bb26dcd9ac75c7c7b74f7ff71d0bfc80" Sep 30 14:36:23 crc kubenswrapper[4840]: I0930 14:36:23.062323 4840 scope.go:117] "RemoveContainer" containerID="e1e2bfc7f9537971d768f1b7183216797522532290c5822edafa8e0b5232aa98" Sep 30 14:36:23 crc kubenswrapper[4840]: E0930 14:36:23.062722 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:36:30 crc kubenswrapper[4840]: I0930 14:36:30.573846 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tql84"] Sep 30 14:36:30 crc kubenswrapper[4840]: I0930 14:36:30.577079 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tql84" Sep 30 14:36:30 crc kubenswrapper[4840]: I0930 14:36:30.586295 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tql84"] Sep 30 14:36:30 crc kubenswrapper[4840]: I0930 14:36:30.674847 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7de6123-432f-4eb4-9814-f017269fadc3-catalog-content\") pod \"redhat-marketplace-tql84\" (UID: \"f7de6123-432f-4eb4-9814-f017269fadc3\") " pod="openshift-marketplace/redhat-marketplace-tql84" Sep 30 14:36:30 crc kubenswrapper[4840]: I0930 14:36:30.675168 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7de6123-432f-4eb4-9814-f017269fadc3-utilities\") pod \"redhat-marketplace-tql84\" (UID: \"f7de6123-432f-4eb4-9814-f017269fadc3\") " pod="openshift-marketplace/redhat-marketplace-tql84" Sep 30 14:36:30 crc kubenswrapper[4840]: I0930 14:36:30.675236 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqmlk\" (UniqueName: \"kubernetes.io/projected/f7de6123-432f-4eb4-9814-f017269fadc3-kube-api-access-kqmlk\") pod \"redhat-marketplace-tql84\" (UID: \"f7de6123-432f-4eb4-9814-f017269fadc3\") " pod="openshift-marketplace/redhat-marketplace-tql84" Sep 30 14:36:30 crc kubenswrapper[4840]: I0930 14:36:30.776500 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7de6123-432f-4eb4-9814-f017269fadc3-utilities\") pod \"redhat-marketplace-tql84\" (UID: \"f7de6123-432f-4eb4-9814-f017269fadc3\") " pod="openshift-marketplace/redhat-marketplace-tql84" Sep 30 14:36:30 crc kubenswrapper[4840]: I0930 14:36:30.776563 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqmlk\" (UniqueName: \"kubernetes.io/projected/f7de6123-432f-4eb4-9814-f017269fadc3-kube-api-access-kqmlk\") pod \"redhat-marketplace-tql84\" (UID: \"f7de6123-432f-4eb4-9814-f017269fadc3\") " pod="openshift-marketplace/redhat-marketplace-tql84" Sep 30 14:36:30 crc kubenswrapper[4840]: I0930 14:36:30.776623 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7de6123-432f-4eb4-9814-f017269fadc3-catalog-content\") pod \"redhat-marketplace-tql84\" (UID: \"f7de6123-432f-4eb4-9814-f017269fadc3\") " pod="openshift-marketplace/redhat-marketplace-tql84" Sep 30 14:36:30 crc kubenswrapper[4840]: I0930 14:36:30.776982 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7de6123-432f-4eb4-9814-f017269fadc3-utilities\") pod \"redhat-marketplace-tql84\" (UID: \"f7de6123-432f-4eb4-9814-f017269fadc3\") " pod="openshift-marketplace/redhat-marketplace-tql84" Sep 30 14:36:30 crc kubenswrapper[4840]: I0930 14:36:30.776987 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7de6123-432f-4eb4-9814-f017269fadc3-catalog-content\") pod \"redhat-marketplace-tql84\" (UID: \"f7de6123-432f-4eb4-9814-f017269fadc3\") " pod="openshift-marketplace/redhat-marketplace-tql84" Sep 30 14:36:30 crc kubenswrapper[4840]: I0930 14:36:30.808994 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqmlk\" (UniqueName: \"kubernetes.io/projected/f7de6123-432f-4eb4-9814-f017269fadc3-kube-api-access-kqmlk\") pod \"redhat-marketplace-tql84\" (UID: \"f7de6123-432f-4eb4-9814-f017269fadc3\") " pod="openshift-marketplace/redhat-marketplace-tql84" Sep 30 14:36:30 crc kubenswrapper[4840]: I0930 14:36:30.899652 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tql84" Sep 30 14:36:31 crc kubenswrapper[4840]: I0930 14:36:31.358937 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tql84"] Sep 30 14:36:32 crc kubenswrapper[4840]: I0930 14:36:32.134302 4840 generic.go:334] "Generic (PLEG): container finished" podID="f7de6123-432f-4eb4-9814-f017269fadc3" containerID="7f3bcb3e69b5245615daa985639116860c32d4a4259f7a8101f3aadb9e2eb59e" exitCode=0 Sep 30 14:36:32 crc kubenswrapper[4840]: I0930 14:36:32.134362 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tql84" event={"ID":"f7de6123-432f-4eb4-9814-f017269fadc3","Type":"ContainerDied","Data":"7f3bcb3e69b5245615daa985639116860c32d4a4259f7a8101f3aadb9e2eb59e"} Sep 30 14:36:32 crc kubenswrapper[4840]: I0930 14:36:32.134590 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tql84" event={"ID":"f7de6123-432f-4eb4-9814-f017269fadc3","Type":"ContainerStarted","Data":"097af74d7aa494848e0fe6a5b2656f194eaa5834877abf802b0fff8c1f6c9b89"} Sep 30 14:36:33 crc kubenswrapper[4840]: I0930 14:36:33.146686 4840 generic.go:334] "Generic (PLEG): container finished" podID="f7de6123-432f-4eb4-9814-f017269fadc3" containerID="0149a23e675633649de7272b51e997ae039e18977161c3ea1fbbca37490886a7" exitCode=0 Sep 30 14:36:33 crc kubenswrapper[4840]: I0930 14:36:33.146754 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tql84" event={"ID":"f7de6123-432f-4eb4-9814-f017269fadc3","Type":"ContainerDied","Data":"0149a23e675633649de7272b51e997ae039e18977161c3ea1fbbca37490886a7"} Sep 30 14:36:34 crc kubenswrapper[4840]: I0930 14:36:34.117321 4840 scope.go:117] "RemoveContainer" containerID="e1e2bfc7f9537971d768f1b7183216797522532290c5822edafa8e0b5232aa98" Sep 30 14:36:34 crc kubenswrapper[4840]: E0930 14:36:34.117919 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:36:34 crc kubenswrapper[4840]: I0930 14:36:34.156919 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tql84" event={"ID":"f7de6123-432f-4eb4-9814-f017269fadc3","Type":"ContainerStarted","Data":"6c74d4a99f30b8a4921dd83b3cc561a3210a56a1405c7aeeedff3a6f5c3b310a"} Sep 30 14:36:34 crc kubenswrapper[4840]: I0930 14:36:34.177366 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tql84" podStartSLOduration=2.695833565 podStartE2EDuration="4.17734823s" podCreationTimestamp="2025-09-30 14:36:30 +0000 UTC" firstStartedPulling="2025-09-30 14:36:32.13606725 +0000 UTC m=+2420.765153663" lastFinishedPulling="2025-09-30 14:36:33.617581905 +0000 UTC m=+2422.246668328" observedRunningTime="2025-09-30 14:36:34.172701424 +0000 UTC m=+2422.801787857" watchObservedRunningTime="2025-09-30 14:36:34.17734823 +0000 UTC m=+2422.806434653" Sep 30 14:36:40 crc kubenswrapper[4840]: I0930 14:36:40.900736 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tql84" Sep 30 14:36:40 crc kubenswrapper[4840]: I0930 14:36:40.901355 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tql84" Sep 30 14:36:40 crc kubenswrapper[4840]: I0930 14:36:40.942758 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tql84" Sep 30 14:36:41 crc kubenswrapper[4840]: I0930 14:36:41.280059 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tql84" Sep 30 14:36:41 crc kubenswrapper[4840]: I0930 14:36:41.327610 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tql84"] Sep 30 14:36:43 crc kubenswrapper[4840]: I0930 14:36:43.245652 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tql84" podUID="f7de6123-432f-4eb4-9814-f017269fadc3" containerName="registry-server" containerID="cri-o://6c74d4a99f30b8a4921dd83b3cc561a3210a56a1405c7aeeedff3a6f5c3b310a" gracePeriod=2 Sep 30 14:36:43 crc kubenswrapper[4840]: I0930 14:36:43.669217 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tql84" Sep 30 14:36:43 crc kubenswrapper[4840]: I0930 14:36:43.724491 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7de6123-432f-4eb4-9814-f017269fadc3-catalog-content\") pod \"f7de6123-432f-4eb4-9814-f017269fadc3\" (UID: \"f7de6123-432f-4eb4-9814-f017269fadc3\") " Sep 30 14:36:43 crc kubenswrapper[4840]: I0930 14:36:43.724750 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqmlk\" (UniqueName: \"kubernetes.io/projected/f7de6123-432f-4eb4-9814-f017269fadc3-kube-api-access-kqmlk\") pod \"f7de6123-432f-4eb4-9814-f017269fadc3\" (UID: \"f7de6123-432f-4eb4-9814-f017269fadc3\") " Sep 30 14:36:43 crc kubenswrapper[4840]: I0930 14:36:43.725047 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7de6123-432f-4eb4-9814-f017269fadc3-utilities\") pod \"f7de6123-432f-4eb4-9814-f017269fadc3\" (UID: \"f7de6123-432f-4eb4-9814-f017269fadc3\") " Sep 30 14:36:43 crc kubenswrapper[4840]: I0930 14:36:43.726248 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7de6123-432f-4eb4-9814-f017269fadc3-utilities" (OuterVolumeSpecName: "utilities") pod "f7de6123-432f-4eb4-9814-f017269fadc3" (UID: "f7de6123-432f-4eb4-9814-f017269fadc3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:36:43 crc kubenswrapper[4840]: I0930 14:36:43.733338 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7de6123-432f-4eb4-9814-f017269fadc3-kube-api-access-kqmlk" (OuterVolumeSpecName: "kube-api-access-kqmlk") pod "f7de6123-432f-4eb4-9814-f017269fadc3" (UID: "f7de6123-432f-4eb4-9814-f017269fadc3"). InnerVolumeSpecName "kube-api-access-kqmlk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:36:43 crc kubenswrapper[4840]: I0930 14:36:43.738995 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7de6123-432f-4eb4-9814-f017269fadc3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f7de6123-432f-4eb4-9814-f017269fadc3" (UID: "f7de6123-432f-4eb4-9814-f017269fadc3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:36:43 crc kubenswrapper[4840]: I0930 14:36:43.827260 4840 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7de6123-432f-4eb4-9814-f017269fadc3-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:36:43 crc kubenswrapper[4840]: I0930 14:36:43.827300 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqmlk\" (UniqueName: \"kubernetes.io/projected/f7de6123-432f-4eb4-9814-f017269fadc3-kube-api-access-kqmlk\") on node \"crc\" DevicePath \"\"" Sep 30 14:36:43 crc kubenswrapper[4840]: I0930 14:36:43.827313 4840 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7de6123-432f-4eb4-9814-f017269fadc3-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:36:44 crc kubenswrapper[4840]: I0930 14:36:44.260046 4840 generic.go:334] "Generic (PLEG): container finished" podID="f7de6123-432f-4eb4-9814-f017269fadc3" containerID="6c74d4a99f30b8a4921dd83b3cc561a3210a56a1405c7aeeedff3a6f5c3b310a" exitCode=0 Sep 30 14:36:44 crc kubenswrapper[4840]: I0930 14:36:44.260093 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tql84" event={"ID":"f7de6123-432f-4eb4-9814-f017269fadc3","Type":"ContainerDied","Data":"6c74d4a99f30b8a4921dd83b3cc561a3210a56a1405c7aeeedff3a6f5c3b310a"} Sep 30 14:36:44 crc kubenswrapper[4840]: I0930 14:36:44.260132 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tql84" event={"ID":"f7de6123-432f-4eb4-9814-f017269fadc3","Type":"ContainerDied","Data":"097af74d7aa494848e0fe6a5b2656f194eaa5834877abf802b0fff8c1f6c9b89"} Sep 30 14:36:44 crc kubenswrapper[4840]: I0930 14:36:44.260141 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tql84" Sep 30 14:36:44 crc kubenswrapper[4840]: I0930 14:36:44.260164 4840 scope.go:117] "RemoveContainer" containerID="6c74d4a99f30b8a4921dd83b3cc561a3210a56a1405c7aeeedff3a6f5c3b310a" Sep 30 14:36:44 crc kubenswrapper[4840]: I0930 14:36:44.283566 4840 scope.go:117] "RemoveContainer" containerID="0149a23e675633649de7272b51e997ae039e18977161c3ea1fbbca37490886a7" Sep 30 14:36:44 crc kubenswrapper[4840]: I0930 14:36:44.284201 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tql84"] Sep 30 14:36:44 crc kubenswrapper[4840]: I0930 14:36:44.291910 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tql84"] Sep 30 14:36:44 crc kubenswrapper[4840]: I0930 14:36:44.303036 4840 scope.go:117] "RemoveContainer" containerID="7f3bcb3e69b5245615daa985639116860c32d4a4259f7a8101f3aadb9e2eb59e" Sep 30 14:36:44 crc kubenswrapper[4840]: I0930 14:36:44.347846 4840 scope.go:117] "RemoveContainer" containerID="6c74d4a99f30b8a4921dd83b3cc561a3210a56a1405c7aeeedff3a6f5c3b310a" Sep 30 14:36:44 crc kubenswrapper[4840]: E0930 14:36:44.348305 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c74d4a99f30b8a4921dd83b3cc561a3210a56a1405c7aeeedff3a6f5c3b310a\": container with ID starting with 6c74d4a99f30b8a4921dd83b3cc561a3210a56a1405c7aeeedff3a6f5c3b310a not found: ID does not exist" containerID="6c74d4a99f30b8a4921dd83b3cc561a3210a56a1405c7aeeedff3a6f5c3b310a" Sep 30 14:36:44 crc kubenswrapper[4840]: I0930 14:36:44.348370 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c74d4a99f30b8a4921dd83b3cc561a3210a56a1405c7aeeedff3a6f5c3b310a"} err="failed to get container status \"6c74d4a99f30b8a4921dd83b3cc561a3210a56a1405c7aeeedff3a6f5c3b310a\": rpc error: code = NotFound desc = could not find container \"6c74d4a99f30b8a4921dd83b3cc561a3210a56a1405c7aeeedff3a6f5c3b310a\": container with ID starting with 6c74d4a99f30b8a4921dd83b3cc561a3210a56a1405c7aeeedff3a6f5c3b310a not found: ID does not exist" Sep 30 14:36:44 crc kubenswrapper[4840]: I0930 14:36:44.348406 4840 scope.go:117] "RemoveContainer" containerID="0149a23e675633649de7272b51e997ae039e18977161c3ea1fbbca37490886a7" Sep 30 14:36:44 crc kubenswrapper[4840]: E0930 14:36:44.348797 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0149a23e675633649de7272b51e997ae039e18977161c3ea1fbbca37490886a7\": container with ID starting with 0149a23e675633649de7272b51e997ae039e18977161c3ea1fbbca37490886a7 not found: ID does not exist" containerID="0149a23e675633649de7272b51e997ae039e18977161c3ea1fbbca37490886a7" Sep 30 14:36:44 crc kubenswrapper[4840]: I0930 14:36:44.348839 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0149a23e675633649de7272b51e997ae039e18977161c3ea1fbbca37490886a7"} err="failed to get container status \"0149a23e675633649de7272b51e997ae039e18977161c3ea1fbbca37490886a7\": rpc error: code = NotFound desc = could not find container \"0149a23e675633649de7272b51e997ae039e18977161c3ea1fbbca37490886a7\": container with ID starting with 0149a23e675633649de7272b51e997ae039e18977161c3ea1fbbca37490886a7 not found: ID does not exist" Sep 30 14:36:44 crc kubenswrapper[4840]: I0930 14:36:44.348856 4840 scope.go:117] "RemoveContainer" containerID="7f3bcb3e69b5245615daa985639116860c32d4a4259f7a8101f3aadb9e2eb59e" Sep 30 14:36:44 crc kubenswrapper[4840]: E0930 14:36:44.349117 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f3bcb3e69b5245615daa985639116860c32d4a4259f7a8101f3aadb9e2eb59e\": container with ID starting with 7f3bcb3e69b5245615daa985639116860c32d4a4259f7a8101f3aadb9e2eb59e not found: ID does not exist" containerID="7f3bcb3e69b5245615daa985639116860c32d4a4259f7a8101f3aadb9e2eb59e" Sep 30 14:36:44 crc kubenswrapper[4840]: I0930 14:36:44.349149 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f3bcb3e69b5245615daa985639116860c32d4a4259f7a8101f3aadb9e2eb59e"} err="failed to get container status \"7f3bcb3e69b5245615daa985639116860c32d4a4259f7a8101f3aadb9e2eb59e\": rpc error: code = NotFound desc = could not find container \"7f3bcb3e69b5245615daa985639116860c32d4a4259f7a8101f3aadb9e2eb59e\": container with ID starting with 7f3bcb3e69b5245615daa985639116860c32d4a4259f7a8101f3aadb9e2eb59e not found: ID does not exist" Sep 30 14:36:45 crc kubenswrapper[4840]: I0930 14:36:45.116383 4840 scope.go:117] "RemoveContainer" containerID="e1e2bfc7f9537971d768f1b7183216797522532290c5822edafa8e0b5232aa98" Sep 30 14:36:45 crc kubenswrapper[4840]: E0930 14:36:45.117151 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:36:46 crc kubenswrapper[4840]: I0930 14:36:46.130004 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7de6123-432f-4eb4-9814-f017269fadc3" path="/var/lib/kubelet/pods/f7de6123-432f-4eb4-9814-f017269fadc3/volumes" Sep 30 14:36:57 crc kubenswrapper[4840]: I0930 14:36:57.117527 4840 scope.go:117] "RemoveContainer" containerID="e1e2bfc7f9537971d768f1b7183216797522532290c5822edafa8e0b5232aa98" Sep 30 14:36:57 crc kubenswrapper[4840]: E0930 14:36:57.118247 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:37:00 crc kubenswrapper[4840]: I0930 14:37:00.407972 4840 generic.go:334] "Generic (PLEG): container finished" podID="20b6236d-9397-4095-a35a-6c24222b852f" containerID="3abfd0972e294f9859f00dc0f8bf56da70b63dd26a76e86e1793f33b0d90451a" exitCode=0 Sep 30 14:37:00 crc kubenswrapper[4840]: I0930 14:37:00.408057 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5q27c" event={"ID":"20b6236d-9397-4095-a35a-6c24222b852f","Type":"ContainerDied","Data":"3abfd0972e294f9859f00dc0f8bf56da70b63dd26a76e86e1793f33b0d90451a"} Sep 30 14:37:01 crc kubenswrapper[4840]: I0930 14:37:01.817434 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5q27c" Sep 30 14:37:01 crc kubenswrapper[4840]: I0930 14:37:01.872933 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-ssh-key\") pod \"20b6236d-9397-4095-a35a-6c24222b852f\" (UID: \"20b6236d-9397-4095-a35a-6c24222b852f\") " Sep 30 14:37:01 crc kubenswrapper[4840]: I0930 14:37:01.873039 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-inventory\") pod \"20b6236d-9397-4095-a35a-6c24222b852f\" (UID: \"20b6236d-9397-4095-a35a-6c24222b852f\") " Sep 30 14:37:01 crc kubenswrapper[4840]: I0930 14:37:01.873097 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-nova-migration-ssh-key-1\") pod \"20b6236d-9397-4095-a35a-6c24222b852f\" (UID: \"20b6236d-9397-4095-a35a-6c24222b852f\") " Sep 30 14:37:01 crc kubenswrapper[4840]: I0930 14:37:01.873149 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-nova-cell1-compute-config-1\") pod \"20b6236d-9397-4095-a35a-6c24222b852f\" (UID: \"20b6236d-9397-4095-a35a-6c24222b852f\") " Sep 30 14:37:01 crc kubenswrapper[4840]: I0930 14:37:01.873264 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/20b6236d-9397-4095-a35a-6c24222b852f-nova-extra-config-0\") pod \"20b6236d-9397-4095-a35a-6c24222b852f\" (UID: \"20b6236d-9397-4095-a35a-6c24222b852f\") " Sep 30 14:37:01 crc kubenswrapper[4840]: I0930 14:37:01.873313 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrmc9\" (UniqueName: \"kubernetes.io/projected/20b6236d-9397-4095-a35a-6c24222b852f-kube-api-access-qrmc9\") pod \"20b6236d-9397-4095-a35a-6c24222b852f\" (UID: \"20b6236d-9397-4095-a35a-6c24222b852f\") " Sep 30 14:37:01 crc kubenswrapper[4840]: I0930 14:37:01.873358 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-nova-cell1-compute-config-0\") pod \"20b6236d-9397-4095-a35a-6c24222b852f\" (UID: \"20b6236d-9397-4095-a35a-6c24222b852f\") " Sep 30 14:37:01 crc kubenswrapper[4840]: I0930 14:37:01.873380 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-nova-migration-ssh-key-0\") pod \"20b6236d-9397-4095-a35a-6c24222b852f\" (UID: \"20b6236d-9397-4095-a35a-6c24222b852f\") " Sep 30 14:37:01 crc kubenswrapper[4840]: I0930 14:37:01.873436 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-nova-combined-ca-bundle\") pod \"20b6236d-9397-4095-a35a-6c24222b852f\" (UID: \"20b6236d-9397-4095-a35a-6c24222b852f\") " Sep 30 14:37:01 crc kubenswrapper[4840]: I0930 14:37:01.880338 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "20b6236d-9397-4095-a35a-6c24222b852f" (UID: "20b6236d-9397-4095-a35a-6c24222b852f"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:37:01 crc kubenswrapper[4840]: I0930 14:37:01.881799 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b6236d-9397-4095-a35a-6c24222b852f-kube-api-access-qrmc9" (OuterVolumeSpecName: "kube-api-access-qrmc9") pod "20b6236d-9397-4095-a35a-6c24222b852f" (UID: "20b6236d-9397-4095-a35a-6c24222b852f"). InnerVolumeSpecName "kube-api-access-qrmc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:37:01 crc kubenswrapper[4840]: I0930 14:37:01.901129 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-inventory" (OuterVolumeSpecName: "inventory") pod "20b6236d-9397-4095-a35a-6c24222b852f" (UID: "20b6236d-9397-4095-a35a-6c24222b852f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:37:01 crc kubenswrapper[4840]: I0930 14:37:01.905054 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "20b6236d-9397-4095-a35a-6c24222b852f" (UID: "20b6236d-9397-4095-a35a-6c24222b852f"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:37:01 crc kubenswrapper[4840]: I0930 14:37:01.908895 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "20b6236d-9397-4095-a35a-6c24222b852f" (UID: "20b6236d-9397-4095-a35a-6c24222b852f"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:37:01 crc kubenswrapper[4840]: I0930 14:37:01.910487 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "20b6236d-9397-4095-a35a-6c24222b852f" (UID: "20b6236d-9397-4095-a35a-6c24222b852f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:37:01 crc kubenswrapper[4840]: I0930 14:37:01.912579 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "20b6236d-9397-4095-a35a-6c24222b852f" (UID: "20b6236d-9397-4095-a35a-6c24222b852f"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:37:01 crc kubenswrapper[4840]: I0930 14:37:01.916649 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "20b6236d-9397-4095-a35a-6c24222b852f" (UID: "20b6236d-9397-4095-a35a-6c24222b852f"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:37:01 crc kubenswrapper[4840]: I0930 14:37:01.920144 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20b6236d-9397-4095-a35a-6c24222b852f-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "20b6236d-9397-4095-a35a-6c24222b852f" (UID: "20b6236d-9397-4095-a35a-6c24222b852f"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:37:01 crc kubenswrapper[4840]: I0930 14:37:01.975098 4840 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:37:01 crc kubenswrapper[4840]: I0930 14:37:01.975137 4840 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 14:37:01 crc kubenswrapper[4840]: I0930 14:37:01.975151 4840 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Sep 30 14:37:01 crc kubenswrapper[4840]: I0930 14:37:01.975164 4840 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Sep 30 14:37:01 crc kubenswrapper[4840]: I0930 14:37:01.975177 4840 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/20b6236d-9397-4095-a35a-6c24222b852f-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:37:01 crc kubenswrapper[4840]: I0930 14:37:01.975188 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrmc9\" (UniqueName: \"kubernetes.io/projected/20b6236d-9397-4095-a35a-6c24222b852f-kube-api-access-qrmc9\") on node \"crc\" DevicePath \"\"" Sep 30 14:37:01 crc kubenswrapper[4840]: I0930 14:37:01.975199 4840 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:37:01 crc kubenswrapper[4840]: I0930 14:37:01.975210 4840 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:37:01 crc kubenswrapper[4840]: I0930 14:37:01.975225 4840 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20b6236d-9397-4095-a35a-6c24222b852f-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:37:02 crc kubenswrapper[4840]: I0930 14:37:02.426830 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5q27c" event={"ID":"20b6236d-9397-4095-a35a-6c24222b852f","Type":"ContainerDied","Data":"4dc73b0b5dae9d180187b29a11f347b8ca9690183b66cb6f3e46008d2815475c"} Sep 30 14:37:02 crc kubenswrapper[4840]: I0930 14:37:02.426873 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4dc73b0b5dae9d180187b29a11f347b8ca9690183b66cb6f3e46008d2815475c" Sep 30 14:37:02 crc kubenswrapper[4840]: I0930 14:37:02.426897 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5q27c" Sep 30 14:37:02 crc kubenswrapper[4840]: I0930 14:37:02.534479 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dzddj"] Sep 30 14:37:02 crc kubenswrapper[4840]: E0930 14:37:02.536383 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7de6123-432f-4eb4-9814-f017269fadc3" containerName="extract-utilities" Sep 30 14:37:02 crc kubenswrapper[4840]: I0930 14:37:02.536429 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7de6123-432f-4eb4-9814-f017269fadc3" containerName="extract-utilities" Sep 30 14:37:02 crc kubenswrapper[4840]: E0930 14:37:02.536473 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20b6236d-9397-4095-a35a-6c24222b852f" containerName="nova-edpm-deployment-openstack-edpm-ipam" Sep 30 14:37:02 crc kubenswrapper[4840]: I0930 14:37:02.536481 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="20b6236d-9397-4095-a35a-6c24222b852f" containerName="nova-edpm-deployment-openstack-edpm-ipam" Sep 30 14:37:02 crc kubenswrapper[4840]: E0930 14:37:02.536593 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7de6123-432f-4eb4-9814-f017269fadc3" containerName="registry-server" Sep 30 14:37:02 crc kubenswrapper[4840]: I0930 14:37:02.536604 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7de6123-432f-4eb4-9814-f017269fadc3" containerName="registry-server" Sep 30 14:37:02 crc kubenswrapper[4840]: E0930 14:37:02.536633 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7de6123-432f-4eb4-9814-f017269fadc3" containerName="extract-content" Sep 30 14:37:02 crc kubenswrapper[4840]: I0930 14:37:02.536690 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7de6123-432f-4eb4-9814-f017269fadc3" containerName="extract-content" Sep 30 14:37:02 crc kubenswrapper[4840]: I0930 14:37:02.539509 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7de6123-432f-4eb4-9814-f017269fadc3" containerName="registry-server" Sep 30 14:37:02 crc kubenswrapper[4840]: I0930 14:37:02.539546 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="20b6236d-9397-4095-a35a-6c24222b852f" containerName="nova-edpm-deployment-openstack-edpm-ipam" Sep 30 14:37:02 crc kubenswrapper[4840]: I0930 14:37:02.540599 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dzddj" Sep 30 14:37:02 crc kubenswrapper[4840]: I0930 14:37:02.543600 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Sep 30 14:37:02 crc kubenswrapper[4840]: I0930 14:37:02.543853 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9zm9z" Sep 30 14:37:02 crc kubenswrapper[4840]: I0930 14:37:02.543963 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 14:37:02 crc kubenswrapper[4840]: I0930 14:37:02.543868 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 14:37:02 crc kubenswrapper[4840]: I0930 14:37:02.548921 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 14:37:02 crc kubenswrapper[4840]: I0930 14:37:02.550071 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dzddj"] Sep 30 14:37:02 crc kubenswrapper[4840]: I0930 14:37:02.583850 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/73a8aa6c-814e-420f-bb1d-666259304a7e-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dzddj\" (UID: \"73a8aa6c-814e-420f-bb1d-666259304a7e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dzddj" Sep 30 14:37:02 crc kubenswrapper[4840]: I0930 14:37:02.583986 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/73a8aa6c-814e-420f-bb1d-666259304a7e-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dzddj\" (UID: \"73a8aa6c-814e-420f-bb1d-666259304a7e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dzddj" Sep 30 14:37:02 crc kubenswrapper[4840]: I0930 14:37:02.584118 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/73a8aa6c-814e-420f-bb1d-666259304a7e-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dzddj\" (UID: \"73a8aa6c-814e-420f-bb1d-666259304a7e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dzddj" Sep 30 14:37:02 crc kubenswrapper[4840]: I0930 14:37:02.584145 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/73a8aa6c-814e-420f-bb1d-666259304a7e-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dzddj\" (UID: \"73a8aa6c-814e-420f-bb1d-666259304a7e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dzddj" Sep 30 14:37:02 crc kubenswrapper[4840]: I0930 14:37:02.584209 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73a8aa6c-814e-420f-bb1d-666259304a7e-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dzddj\" (UID: \"73a8aa6c-814e-420f-bb1d-666259304a7e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dzddj" Sep 30 14:37:02 crc kubenswrapper[4840]: I0930 14:37:02.584243 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrfkh\" (UniqueName: \"kubernetes.io/projected/73a8aa6c-814e-420f-bb1d-666259304a7e-kube-api-access-nrfkh\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dzddj\" (UID: \"73a8aa6c-814e-420f-bb1d-666259304a7e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dzddj" Sep 30 14:37:02 crc kubenswrapper[4840]: I0930 14:37:02.584271 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/73a8aa6c-814e-420f-bb1d-666259304a7e-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dzddj\" (UID: \"73a8aa6c-814e-420f-bb1d-666259304a7e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dzddj" Sep 30 14:37:02 crc kubenswrapper[4840]: I0930 14:37:02.685493 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/73a8aa6c-814e-420f-bb1d-666259304a7e-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dzddj\" (UID: \"73a8aa6c-814e-420f-bb1d-666259304a7e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dzddj" Sep 30 14:37:02 crc kubenswrapper[4840]: I0930 14:37:02.685896 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/73a8aa6c-814e-420f-bb1d-666259304a7e-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dzddj\" (UID: \"73a8aa6c-814e-420f-bb1d-666259304a7e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dzddj" Sep 30 14:37:02 crc kubenswrapper[4840]: I0930 14:37:02.685959 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/73a8aa6c-814e-420f-bb1d-666259304a7e-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dzddj\" (UID: \"73a8aa6c-814e-420f-bb1d-666259304a7e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dzddj" Sep 30 14:37:02 crc kubenswrapper[4840]: I0930 14:37:02.685984 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/73a8aa6c-814e-420f-bb1d-666259304a7e-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dzddj\" (UID: \"73a8aa6c-814e-420f-bb1d-666259304a7e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dzddj" Sep 30 14:37:02 crc kubenswrapper[4840]: I0930 14:37:02.686013 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73a8aa6c-814e-420f-bb1d-666259304a7e-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dzddj\" (UID: \"73a8aa6c-814e-420f-bb1d-666259304a7e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dzddj" Sep 30 14:37:02 crc kubenswrapper[4840]: I0930 14:37:02.686136 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrfkh\" (UniqueName: \"kubernetes.io/projected/73a8aa6c-814e-420f-bb1d-666259304a7e-kube-api-access-nrfkh\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dzddj\" (UID: \"73a8aa6c-814e-420f-bb1d-666259304a7e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dzddj" Sep 30 14:37:02 crc kubenswrapper[4840]: I0930 14:37:02.686532 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/73a8aa6c-814e-420f-bb1d-666259304a7e-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dzddj\" (UID: \"73a8aa6c-814e-420f-bb1d-666259304a7e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dzddj" Sep 30 14:37:02 crc kubenswrapper[4840]: I0930 14:37:02.689436 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/73a8aa6c-814e-420f-bb1d-666259304a7e-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dzddj\" (UID: \"73a8aa6c-814e-420f-bb1d-666259304a7e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dzddj" Sep 30 14:37:02 crc kubenswrapper[4840]: I0930 14:37:02.689436 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/73a8aa6c-814e-420f-bb1d-666259304a7e-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dzddj\" (UID: \"73a8aa6c-814e-420f-bb1d-666259304a7e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dzddj" Sep 30 14:37:02 crc kubenswrapper[4840]: I0930 14:37:02.689653 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/73a8aa6c-814e-420f-bb1d-666259304a7e-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dzddj\" (UID: \"73a8aa6c-814e-420f-bb1d-666259304a7e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dzddj" Sep 30 14:37:02 crc kubenswrapper[4840]: I0930 14:37:02.690108 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/73a8aa6c-814e-420f-bb1d-666259304a7e-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dzddj\" (UID: \"73a8aa6c-814e-420f-bb1d-666259304a7e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dzddj" Sep 30 14:37:02 crc kubenswrapper[4840]: I0930 14:37:02.691173 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73a8aa6c-814e-420f-bb1d-666259304a7e-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dzddj\" (UID: \"73a8aa6c-814e-420f-bb1d-666259304a7e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dzddj" Sep 30 14:37:02 crc kubenswrapper[4840]: I0930 14:37:02.692440 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/73a8aa6c-814e-420f-bb1d-666259304a7e-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dzddj\" (UID: \"73a8aa6c-814e-420f-bb1d-666259304a7e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dzddj" Sep 30 14:37:02 crc kubenswrapper[4840]: I0930 14:37:02.705129 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrfkh\" (UniqueName: \"kubernetes.io/projected/73a8aa6c-814e-420f-bb1d-666259304a7e-kube-api-access-nrfkh\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dzddj\" (UID: \"73a8aa6c-814e-420f-bb1d-666259304a7e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dzddj" Sep 30 14:37:02 crc kubenswrapper[4840]: I0930 14:37:02.859716 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dzddj" Sep 30 14:37:03 crc kubenswrapper[4840]: I0930 14:37:03.360004 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dzddj"] Sep 30 14:37:03 crc kubenswrapper[4840]: I0930 14:37:03.436161 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dzddj" event={"ID":"73a8aa6c-814e-420f-bb1d-666259304a7e","Type":"ContainerStarted","Data":"7b4c4354ad6598d862fbce8cbac85d37ffe9d565ba89b9439f6fb0bd36d811f8"} Sep 30 14:37:04 crc kubenswrapper[4840]: I0930 14:37:04.446012 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dzddj" event={"ID":"73a8aa6c-814e-420f-bb1d-666259304a7e","Type":"ContainerStarted","Data":"44ebca09b8bed4808006027c4565481ad0f28790827f212b172bf15c6f855433"} Sep 30 14:37:04 crc kubenswrapper[4840]: I0930 14:37:04.468250 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dzddj" podStartSLOduration=2.018319703 podStartE2EDuration="2.468226561s" podCreationTimestamp="2025-09-30 14:37:02 +0000 UTC" firstStartedPulling="2025-09-30 14:37:03.364132887 +0000 UTC m=+2451.993219310" lastFinishedPulling="2025-09-30 14:37:03.814039745 +0000 UTC m=+2452.443126168" observedRunningTime="2025-09-30 14:37:04.462716124 +0000 UTC m=+2453.091802547" watchObservedRunningTime="2025-09-30 14:37:04.468226561 +0000 UTC m=+2453.097312984" Sep 30 14:37:09 crc kubenswrapper[4840]: I0930 14:37:09.116865 4840 scope.go:117] "RemoveContainer" containerID="e1e2bfc7f9537971d768f1b7183216797522532290c5822edafa8e0b5232aa98" Sep 30 14:37:09 crc kubenswrapper[4840]: E0930 14:37:09.117740 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:37:24 crc kubenswrapper[4840]: I0930 14:37:24.117446 4840 scope.go:117] "RemoveContainer" containerID="e1e2bfc7f9537971d768f1b7183216797522532290c5822edafa8e0b5232aa98" Sep 30 14:37:24 crc kubenswrapper[4840]: E0930 14:37:24.118337 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:37:37 crc kubenswrapper[4840]: I0930 14:37:37.116355 4840 scope.go:117] "RemoveContainer" containerID="e1e2bfc7f9537971d768f1b7183216797522532290c5822edafa8e0b5232aa98" Sep 30 14:37:37 crc kubenswrapper[4840]: E0930 14:37:37.117138 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:37:49 crc kubenswrapper[4840]: I0930 14:37:49.116750 4840 scope.go:117] "RemoveContainer" containerID="e1e2bfc7f9537971d768f1b7183216797522532290c5822edafa8e0b5232aa98" Sep 30 14:37:49 crc kubenswrapper[4840]: E0930 14:37:49.117538 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:38:02 crc kubenswrapper[4840]: I0930 14:38:02.122494 4840 scope.go:117] "RemoveContainer" containerID="e1e2bfc7f9537971d768f1b7183216797522532290c5822edafa8e0b5232aa98" Sep 30 14:38:02 crc kubenswrapper[4840]: E0930 14:38:02.123338 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:38:14 crc kubenswrapper[4840]: I0930 14:38:14.117123 4840 scope.go:117] "RemoveContainer" containerID="e1e2bfc7f9537971d768f1b7183216797522532290c5822edafa8e0b5232aa98" Sep 30 14:38:14 crc kubenswrapper[4840]: E0930 14:38:14.118149 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:38:25 crc kubenswrapper[4840]: I0930 14:38:25.116727 4840 scope.go:117] "RemoveContainer" containerID="e1e2bfc7f9537971d768f1b7183216797522532290c5822edafa8e0b5232aa98" Sep 30 14:38:25 crc kubenswrapper[4840]: E0930 14:38:25.118350 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:38:37 crc kubenswrapper[4840]: I0930 14:38:37.116629 4840 scope.go:117] "RemoveContainer" containerID="e1e2bfc7f9537971d768f1b7183216797522532290c5822edafa8e0b5232aa98" Sep 30 14:38:37 crc kubenswrapper[4840]: E0930 14:38:37.117429 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:38:48 crc kubenswrapper[4840]: I0930 14:38:48.117158 4840 scope.go:117] "RemoveContainer" containerID="e1e2bfc7f9537971d768f1b7183216797522532290c5822edafa8e0b5232aa98" Sep 30 14:38:48 crc kubenswrapper[4840]: E0930 14:38:48.118055 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:39:03 crc kubenswrapper[4840]: I0930 14:39:03.117313 4840 scope.go:117] "RemoveContainer" containerID="e1e2bfc7f9537971d768f1b7183216797522532290c5822edafa8e0b5232aa98" Sep 30 14:39:03 crc kubenswrapper[4840]: E0930 14:39:03.118161 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:39:14 crc kubenswrapper[4840]: I0930 14:39:14.116464 4840 scope.go:117] "RemoveContainer" containerID="e1e2bfc7f9537971d768f1b7183216797522532290c5822edafa8e0b5232aa98" Sep 30 14:39:14 crc kubenswrapper[4840]: E0930 14:39:14.117640 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:39:23 crc kubenswrapper[4840]: I0930 14:39:23.604489 4840 generic.go:334] "Generic (PLEG): container finished" podID="73a8aa6c-814e-420f-bb1d-666259304a7e" containerID="44ebca09b8bed4808006027c4565481ad0f28790827f212b172bf15c6f855433" exitCode=0 Sep 30 14:39:23 crc kubenswrapper[4840]: I0930 14:39:23.604624 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dzddj" event={"ID":"73a8aa6c-814e-420f-bb1d-666259304a7e","Type":"ContainerDied","Data":"44ebca09b8bed4808006027c4565481ad0f28790827f212b172bf15c6f855433"} Sep 30 14:39:24 crc kubenswrapper[4840]: I0930 14:39:24.992355 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dzddj" Sep 30 14:39:25 crc kubenswrapper[4840]: I0930 14:39:25.117190 4840 scope.go:117] "RemoveContainer" containerID="e1e2bfc7f9537971d768f1b7183216797522532290c5822edafa8e0b5232aa98" Sep 30 14:39:25 crc kubenswrapper[4840]: E0930 14:39:25.117517 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:39:25 crc kubenswrapper[4840]: I0930 14:39:25.127915 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrfkh\" (UniqueName: \"kubernetes.io/projected/73a8aa6c-814e-420f-bb1d-666259304a7e-kube-api-access-nrfkh\") pod \"73a8aa6c-814e-420f-bb1d-666259304a7e\" (UID: \"73a8aa6c-814e-420f-bb1d-666259304a7e\") " Sep 30 14:39:25 crc kubenswrapper[4840]: I0930 14:39:25.128042 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/73a8aa6c-814e-420f-bb1d-666259304a7e-inventory\") pod \"73a8aa6c-814e-420f-bb1d-666259304a7e\" (UID: \"73a8aa6c-814e-420f-bb1d-666259304a7e\") " Sep 30 14:39:25 crc kubenswrapper[4840]: I0930 14:39:25.128111 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/73a8aa6c-814e-420f-bb1d-666259304a7e-ceilometer-compute-config-data-2\") pod \"73a8aa6c-814e-420f-bb1d-666259304a7e\" (UID: \"73a8aa6c-814e-420f-bb1d-666259304a7e\") " Sep 30 14:39:25 crc kubenswrapper[4840]: I0930 14:39:25.128160 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/73a8aa6c-814e-420f-bb1d-666259304a7e-ceilometer-compute-config-data-1\") pod \"73a8aa6c-814e-420f-bb1d-666259304a7e\" (UID: \"73a8aa6c-814e-420f-bb1d-666259304a7e\") " Sep 30 14:39:25 crc kubenswrapper[4840]: I0930 14:39:25.128221 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/73a8aa6c-814e-420f-bb1d-666259304a7e-ceilometer-compute-config-data-0\") pod \"73a8aa6c-814e-420f-bb1d-666259304a7e\" (UID: \"73a8aa6c-814e-420f-bb1d-666259304a7e\") " Sep 30 14:39:25 crc kubenswrapper[4840]: I0930 14:39:25.128244 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73a8aa6c-814e-420f-bb1d-666259304a7e-telemetry-combined-ca-bundle\") pod \"73a8aa6c-814e-420f-bb1d-666259304a7e\" (UID: \"73a8aa6c-814e-420f-bb1d-666259304a7e\") " Sep 30 14:39:25 crc kubenswrapper[4840]: I0930 14:39:25.128305 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/73a8aa6c-814e-420f-bb1d-666259304a7e-ssh-key\") pod \"73a8aa6c-814e-420f-bb1d-666259304a7e\" (UID: \"73a8aa6c-814e-420f-bb1d-666259304a7e\") " Sep 30 14:39:25 crc kubenswrapper[4840]: I0930 14:39:25.133485 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73a8aa6c-814e-420f-bb1d-666259304a7e-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "73a8aa6c-814e-420f-bb1d-666259304a7e" (UID: "73a8aa6c-814e-420f-bb1d-666259304a7e"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:39:25 crc kubenswrapper[4840]: I0930 14:39:25.136896 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73a8aa6c-814e-420f-bb1d-666259304a7e-kube-api-access-nrfkh" (OuterVolumeSpecName: "kube-api-access-nrfkh") pod "73a8aa6c-814e-420f-bb1d-666259304a7e" (UID: "73a8aa6c-814e-420f-bb1d-666259304a7e"). InnerVolumeSpecName "kube-api-access-nrfkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:39:25 crc kubenswrapper[4840]: I0930 14:39:25.161516 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73a8aa6c-814e-420f-bb1d-666259304a7e-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "73a8aa6c-814e-420f-bb1d-666259304a7e" (UID: "73a8aa6c-814e-420f-bb1d-666259304a7e"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:39:25 crc kubenswrapper[4840]: I0930 14:39:25.161957 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73a8aa6c-814e-420f-bb1d-666259304a7e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "73a8aa6c-814e-420f-bb1d-666259304a7e" (UID: "73a8aa6c-814e-420f-bb1d-666259304a7e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:39:25 crc kubenswrapper[4840]: I0930 14:39:25.163217 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73a8aa6c-814e-420f-bb1d-666259304a7e-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "73a8aa6c-814e-420f-bb1d-666259304a7e" (UID: "73a8aa6c-814e-420f-bb1d-666259304a7e"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:39:25 crc kubenswrapper[4840]: I0930 14:39:25.164957 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73a8aa6c-814e-420f-bb1d-666259304a7e-inventory" (OuterVolumeSpecName: "inventory") pod "73a8aa6c-814e-420f-bb1d-666259304a7e" (UID: "73a8aa6c-814e-420f-bb1d-666259304a7e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:39:25 crc kubenswrapper[4840]: I0930 14:39:25.166994 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73a8aa6c-814e-420f-bb1d-666259304a7e-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "73a8aa6c-814e-420f-bb1d-666259304a7e" (UID: "73a8aa6c-814e-420f-bb1d-666259304a7e"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:39:25 crc kubenswrapper[4840]: I0930 14:39:25.231141 4840 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73a8aa6c-814e-420f-bb1d-666259304a7e-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:25 crc kubenswrapper[4840]: I0930 14:39:25.231380 4840 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/73a8aa6c-814e-420f-bb1d-666259304a7e-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:25 crc kubenswrapper[4840]: I0930 14:39:25.231394 4840 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/73a8aa6c-814e-420f-bb1d-666259304a7e-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:25 crc kubenswrapper[4840]: I0930 14:39:25.231410 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrfkh\" (UniqueName: \"kubernetes.io/projected/73a8aa6c-814e-420f-bb1d-666259304a7e-kube-api-access-nrfkh\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:25 crc kubenswrapper[4840]: I0930 14:39:25.231432 4840 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/73a8aa6c-814e-420f-bb1d-666259304a7e-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:25 crc kubenswrapper[4840]: I0930 14:39:25.231450 4840 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/73a8aa6c-814e-420f-bb1d-666259304a7e-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:25 crc kubenswrapper[4840]: I0930 14:39:25.231461 4840 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/73a8aa6c-814e-420f-bb1d-666259304a7e-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:25 crc kubenswrapper[4840]: I0930 14:39:25.627299 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dzddj" event={"ID":"73a8aa6c-814e-420f-bb1d-666259304a7e","Type":"ContainerDied","Data":"7b4c4354ad6598d862fbce8cbac85d37ffe9d565ba89b9439f6fb0bd36d811f8"} Sep 30 14:39:25 crc kubenswrapper[4840]: I0930 14:39:25.627359 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b4c4354ad6598d862fbce8cbac85d37ffe9d565ba89b9439f6fb0bd36d811f8" Sep 30 14:39:25 crc kubenswrapper[4840]: I0930 14:39:25.627394 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dzddj" Sep 30 14:39:36 crc kubenswrapper[4840]: I0930 14:39:36.117385 4840 scope.go:117] "RemoveContainer" containerID="e1e2bfc7f9537971d768f1b7183216797522532290c5822edafa8e0b5232aa98" Sep 30 14:39:36 crc kubenswrapper[4840]: E0930 14:39:36.118006 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:39:47 crc kubenswrapper[4840]: I0930 14:39:47.116536 4840 scope.go:117] "RemoveContainer" containerID="e1e2bfc7f9537971d768f1b7183216797522532290c5822edafa8e0b5232aa98" Sep 30 14:39:47 crc kubenswrapper[4840]: E0930 14:39:47.117588 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:40:00 crc kubenswrapper[4840]: I0930 14:40:00.116544 4840 scope.go:117] "RemoveContainer" containerID="e1e2bfc7f9537971d768f1b7183216797522532290c5822edafa8e0b5232aa98" Sep 30 14:40:00 crc kubenswrapper[4840]: E0930 14:40:00.117476 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:40:09 crc kubenswrapper[4840]: E0930 14:40:09.469296 4840 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.129.56.122:48556->38.129.56.122:37625: write tcp 38.129.56.122:48556->38.129.56.122:37625: write: broken pipe Sep 30 14:40:13 crc kubenswrapper[4840]: I0930 14:40:13.117495 4840 scope.go:117] "RemoveContainer" containerID="e1e2bfc7f9537971d768f1b7183216797522532290c5822edafa8e0b5232aa98" Sep 30 14:40:13 crc kubenswrapper[4840]: E0930 14:40:13.118688 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:40:19 crc kubenswrapper[4840]: I0930 14:40:19.878394 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5jgck"] Sep 30 14:40:19 crc kubenswrapper[4840]: E0930 14:40:19.879436 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73a8aa6c-814e-420f-bb1d-666259304a7e" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Sep 30 14:40:19 crc kubenswrapper[4840]: I0930 14:40:19.879451 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="73a8aa6c-814e-420f-bb1d-666259304a7e" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Sep 30 14:40:19 crc kubenswrapper[4840]: I0930 14:40:19.879665 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="73a8aa6c-814e-420f-bb1d-666259304a7e" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Sep 30 14:40:19 crc kubenswrapper[4840]: I0930 14:40:19.881171 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5jgck" Sep 30 14:40:19 crc kubenswrapper[4840]: I0930 14:40:19.890443 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5jgck"] Sep 30 14:40:19 crc kubenswrapper[4840]: I0930 14:40:19.981330 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8s6ss\" (UniqueName: \"kubernetes.io/projected/858b928e-a0c6-4c2d-9000-9e9807b7aecc-kube-api-access-8s6ss\") pod \"certified-operators-5jgck\" (UID: \"858b928e-a0c6-4c2d-9000-9e9807b7aecc\") " pod="openshift-marketplace/certified-operators-5jgck" Sep 30 14:40:19 crc kubenswrapper[4840]: I0930 14:40:19.981602 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/858b928e-a0c6-4c2d-9000-9e9807b7aecc-catalog-content\") pod \"certified-operators-5jgck\" (UID: \"858b928e-a0c6-4c2d-9000-9e9807b7aecc\") " pod="openshift-marketplace/certified-operators-5jgck" Sep 30 14:40:19 crc kubenswrapper[4840]: I0930 14:40:19.981720 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/858b928e-a0c6-4c2d-9000-9e9807b7aecc-utilities\") pod \"certified-operators-5jgck\" (UID: \"858b928e-a0c6-4c2d-9000-9e9807b7aecc\") " pod="openshift-marketplace/certified-operators-5jgck" Sep 30 14:40:20 crc kubenswrapper[4840]: I0930 14:40:20.083676 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8s6ss\" (UniqueName: \"kubernetes.io/projected/858b928e-a0c6-4c2d-9000-9e9807b7aecc-kube-api-access-8s6ss\") pod \"certified-operators-5jgck\" (UID: \"858b928e-a0c6-4c2d-9000-9e9807b7aecc\") " pod="openshift-marketplace/certified-operators-5jgck" Sep 30 14:40:20 crc kubenswrapper[4840]: I0930 14:40:20.083794 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/858b928e-a0c6-4c2d-9000-9e9807b7aecc-catalog-content\") pod \"certified-operators-5jgck\" (UID: \"858b928e-a0c6-4c2d-9000-9e9807b7aecc\") " pod="openshift-marketplace/certified-operators-5jgck" Sep 30 14:40:20 crc kubenswrapper[4840]: I0930 14:40:20.083848 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/858b928e-a0c6-4c2d-9000-9e9807b7aecc-utilities\") pod \"certified-operators-5jgck\" (UID: \"858b928e-a0c6-4c2d-9000-9e9807b7aecc\") " pod="openshift-marketplace/certified-operators-5jgck" Sep 30 14:40:20 crc kubenswrapper[4840]: I0930 14:40:20.084844 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/858b928e-a0c6-4c2d-9000-9e9807b7aecc-catalog-content\") pod \"certified-operators-5jgck\" (UID: \"858b928e-a0c6-4c2d-9000-9e9807b7aecc\") " pod="openshift-marketplace/certified-operators-5jgck" Sep 30 14:40:20 crc kubenswrapper[4840]: I0930 14:40:20.084892 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/858b928e-a0c6-4c2d-9000-9e9807b7aecc-utilities\") pod \"certified-operators-5jgck\" (UID: \"858b928e-a0c6-4c2d-9000-9e9807b7aecc\") " pod="openshift-marketplace/certified-operators-5jgck" Sep 30 14:40:20 crc kubenswrapper[4840]: I0930 14:40:20.111261 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8s6ss\" (UniqueName: \"kubernetes.io/projected/858b928e-a0c6-4c2d-9000-9e9807b7aecc-kube-api-access-8s6ss\") pod \"certified-operators-5jgck\" (UID: \"858b928e-a0c6-4c2d-9000-9e9807b7aecc\") " pod="openshift-marketplace/certified-operators-5jgck" Sep 30 14:40:20 crc kubenswrapper[4840]: I0930 14:40:20.206080 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5jgck" Sep 30 14:40:20 crc kubenswrapper[4840]: I0930 14:40:20.772071 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5jgck"] Sep 30 14:40:21 crc kubenswrapper[4840]: I0930 14:40:21.148312 4840 generic.go:334] "Generic (PLEG): container finished" podID="858b928e-a0c6-4c2d-9000-9e9807b7aecc" containerID="a27c73cb26c8132a1305dccdf2c5e9543b7f039e17389cf1a5d67a25dac35ab5" exitCode=0 Sep 30 14:40:21 crc kubenswrapper[4840]: I0930 14:40:21.148364 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5jgck" event={"ID":"858b928e-a0c6-4c2d-9000-9e9807b7aecc","Type":"ContainerDied","Data":"a27c73cb26c8132a1305dccdf2c5e9543b7f039e17389cf1a5d67a25dac35ab5"} Sep 30 14:40:21 crc kubenswrapper[4840]: I0930 14:40:21.148398 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5jgck" event={"ID":"858b928e-a0c6-4c2d-9000-9e9807b7aecc","Type":"ContainerStarted","Data":"6db13533a63485c6755720f089cb43bcca34f27cc58a5d24a16c217c2041b8db"} Sep 30 14:40:21 crc kubenswrapper[4840]: I0930 14:40:21.150197 4840 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 14:40:22 crc kubenswrapper[4840]: I0930 14:40:22.159421 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5jgck" event={"ID":"858b928e-a0c6-4c2d-9000-9e9807b7aecc","Type":"ContainerStarted","Data":"4e5391b00e315037cf2ccc869d34d297c823ca74294773b5a1b842141218b6a8"} Sep 30 14:40:23 crc kubenswrapper[4840]: I0930 14:40:23.170361 4840 generic.go:334] "Generic (PLEG): container finished" podID="858b928e-a0c6-4c2d-9000-9e9807b7aecc" containerID="4e5391b00e315037cf2ccc869d34d297c823ca74294773b5a1b842141218b6a8" exitCode=0 Sep 30 14:40:23 crc kubenswrapper[4840]: I0930 14:40:23.170411 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5jgck" event={"ID":"858b928e-a0c6-4c2d-9000-9e9807b7aecc","Type":"ContainerDied","Data":"4e5391b00e315037cf2ccc869d34d297c823ca74294773b5a1b842141218b6a8"} Sep 30 14:40:23 crc kubenswrapper[4840]: I0930 14:40:23.612443 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Sep 30 14:40:23 crc kubenswrapper[4840]: I0930 14:40:23.613987 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Sep 30 14:40:23 crc kubenswrapper[4840]: I0930 14:40:23.616115 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Sep 30 14:40:23 crc kubenswrapper[4840]: I0930 14:40:23.616196 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Sep 30 14:40:23 crc kubenswrapper[4840]: I0930 14:40:23.616510 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-vfxhm" Sep 30 14:40:23 crc kubenswrapper[4840]: I0930 14:40:23.616642 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Sep 30 14:40:23 crc kubenswrapper[4840]: I0930 14:40:23.650042 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Sep 30 14:40:23 crc kubenswrapper[4840]: I0930 14:40:23.754137 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\") " pod="openstack/tempest-tests-tempest" Sep 30 14:40:23 crc kubenswrapper[4840]: I0930 14:40:23.754210 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\") " pod="openstack/tempest-tests-tempest" Sep 30 14:40:23 crc kubenswrapper[4840]: I0930 14:40:23.754294 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-config-data\") pod \"tempest-tests-tempest\" (UID: \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\") " pod="openstack/tempest-tests-tempest" Sep 30 14:40:23 crc kubenswrapper[4840]: I0930 14:40:23.754329 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnfwb\" (UniqueName: \"kubernetes.io/projected/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-kube-api-access-hnfwb\") pod \"tempest-tests-tempest\" (UID: \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\") " pod="openstack/tempest-tests-tempest" Sep 30 14:40:23 crc kubenswrapper[4840]: I0930 14:40:23.754349 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\") " pod="openstack/tempest-tests-tempest" Sep 30 14:40:23 crc kubenswrapper[4840]: I0930 14:40:23.754375 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\") " pod="openstack/tempest-tests-tempest" Sep 30 14:40:23 crc kubenswrapper[4840]: I0930 14:40:23.754415 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\") " pod="openstack/tempest-tests-tempest" Sep 30 14:40:23 crc kubenswrapper[4840]: I0930 14:40:23.754430 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\") " pod="openstack/tempest-tests-tempest" Sep 30 14:40:23 crc kubenswrapper[4840]: I0930 14:40:23.754462 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\") " pod="openstack/tempest-tests-tempest" Sep 30 14:40:23 crc kubenswrapper[4840]: I0930 14:40:23.856108 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-config-data\") pod \"tempest-tests-tempest\" (UID: \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\") " pod="openstack/tempest-tests-tempest" Sep 30 14:40:23 crc kubenswrapper[4840]: I0930 14:40:23.856194 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnfwb\" (UniqueName: \"kubernetes.io/projected/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-kube-api-access-hnfwb\") pod \"tempest-tests-tempest\" (UID: \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\") " pod="openstack/tempest-tests-tempest" Sep 30 14:40:23 crc kubenswrapper[4840]: I0930 14:40:23.856222 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\") " pod="openstack/tempest-tests-tempest" Sep 30 14:40:23 crc kubenswrapper[4840]: I0930 14:40:23.856258 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\") " pod="openstack/tempest-tests-tempest" Sep 30 14:40:23 crc kubenswrapper[4840]: I0930 14:40:23.856346 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\") " pod="openstack/tempest-tests-tempest" Sep 30 14:40:23 crc kubenswrapper[4840]: I0930 14:40:23.856388 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\") " pod="openstack/tempest-tests-tempest" Sep 30 14:40:23 crc kubenswrapper[4840]: I0930 14:40:23.856457 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\") " pod="openstack/tempest-tests-tempest" Sep 30 14:40:23 crc kubenswrapper[4840]: I0930 14:40:23.856575 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\") " pod="openstack/tempest-tests-tempest" Sep 30 14:40:23 crc kubenswrapper[4840]: I0930 14:40:23.856613 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\") " pod="openstack/tempest-tests-tempest" Sep 30 14:40:23 crc kubenswrapper[4840]: I0930 14:40:23.857094 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\") " pod="openstack/tempest-tests-tempest" Sep 30 14:40:23 crc kubenswrapper[4840]: I0930 14:40:23.857206 4840 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/tempest-tests-tempest" Sep 30 14:40:23 crc kubenswrapper[4840]: I0930 14:40:23.857511 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\") " pod="openstack/tempest-tests-tempest" Sep 30 14:40:23 crc kubenswrapper[4840]: I0930 14:40:23.858431 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\") " pod="openstack/tempest-tests-tempest" Sep 30 14:40:23 crc kubenswrapper[4840]: I0930 14:40:23.859623 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-config-data\") pod \"tempest-tests-tempest\" (UID: \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\") " pod="openstack/tempest-tests-tempest" Sep 30 14:40:23 crc kubenswrapper[4840]: I0930 14:40:23.865827 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\") " pod="openstack/tempest-tests-tempest" Sep 30 14:40:23 crc kubenswrapper[4840]: I0930 14:40:23.865936 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\") " pod="openstack/tempest-tests-tempest" Sep 30 14:40:23 crc kubenswrapper[4840]: I0930 14:40:23.866013 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\") " pod="openstack/tempest-tests-tempest" Sep 30 14:40:23 crc kubenswrapper[4840]: I0930 14:40:23.877001 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnfwb\" (UniqueName: \"kubernetes.io/projected/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-kube-api-access-hnfwb\") pod \"tempest-tests-tempest\" (UID: \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\") " pod="openstack/tempest-tests-tempest" Sep 30 14:40:23 crc kubenswrapper[4840]: I0930 14:40:23.892928 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\") " pod="openstack/tempest-tests-tempest" Sep 30 14:40:23 crc kubenswrapper[4840]: I0930 14:40:23.950401 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Sep 30 14:40:24 crc kubenswrapper[4840]: I0930 14:40:24.432941 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Sep 30 14:40:24 crc kubenswrapper[4840]: W0930 14:40:24.436165 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1e0776f6_f8bf_4ef7_a68b_a48fb2e5f545.slice/crio-a4bae75b2bc474b3ecd0feba2bfdebbc43888b95aa8e1cef74f966c86bc5c243 WatchSource:0}: Error finding container a4bae75b2bc474b3ecd0feba2bfdebbc43888b95aa8e1cef74f966c86bc5c243: Status 404 returned error can't find the container with id a4bae75b2bc474b3ecd0feba2bfdebbc43888b95aa8e1cef74f966c86bc5c243 Sep 30 14:40:25 crc kubenswrapper[4840]: I0930 14:40:25.116768 4840 scope.go:117] "RemoveContainer" containerID="e1e2bfc7f9537971d768f1b7183216797522532290c5822edafa8e0b5232aa98" Sep 30 14:40:25 crc kubenswrapper[4840]: E0930 14:40:25.117009 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:40:25 crc kubenswrapper[4840]: I0930 14:40:25.189141 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5jgck" event={"ID":"858b928e-a0c6-4c2d-9000-9e9807b7aecc","Type":"ContainerStarted","Data":"3295b9ab6a6ac3212d6e2f72c2016f956efa124402b317b46d6058bcf548f20e"} Sep 30 14:40:25 crc kubenswrapper[4840]: I0930 14:40:25.190997 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545","Type":"ContainerStarted","Data":"a4bae75b2bc474b3ecd0feba2bfdebbc43888b95aa8e1cef74f966c86bc5c243"} Sep 30 14:40:25 crc kubenswrapper[4840]: I0930 14:40:25.208135 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5jgck" podStartSLOduration=3.343266459 podStartE2EDuration="6.208112746s" podCreationTimestamp="2025-09-30 14:40:19 +0000 UTC" firstStartedPulling="2025-09-30 14:40:21.149975554 +0000 UTC m=+2649.779061977" lastFinishedPulling="2025-09-30 14:40:24.014821841 +0000 UTC m=+2652.643908264" observedRunningTime="2025-09-30 14:40:25.204409567 +0000 UTC m=+2653.833496010" watchObservedRunningTime="2025-09-30 14:40:25.208112746 +0000 UTC m=+2653.837199169" Sep 30 14:40:30 crc kubenswrapper[4840]: I0930 14:40:30.206894 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5jgck" Sep 30 14:40:30 crc kubenswrapper[4840]: I0930 14:40:30.207353 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5jgck" Sep 30 14:40:30 crc kubenswrapper[4840]: I0930 14:40:30.261615 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5jgck" Sep 30 14:40:30 crc kubenswrapper[4840]: I0930 14:40:30.312411 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5jgck" Sep 30 14:40:30 crc kubenswrapper[4840]: I0930 14:40:30.500609 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5jgck"] Sep 30 14:40:32 crc kubenswrapper[4840]: I0930 14:40:32.258117 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5jgck" podUID="858b928e-a0c6-4c2d-9000-9e9807b7aecc" containerName="registry-server" containerID="cri-o://3295b9ab6a6ac3212d6e2f72c2016f956efa124402b317b46d6058bcf548f20e" gracePeriod=2 Sep 30 14:40:33 crc kubenswrapper[4840]: I0930 14:40:33.223234 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5jgck" Sep 30 14:40:33 crc kubenswrapper[4840]: I0930 14:40:33.271194 4840 generic.go:334] "Generic (PLEG): container finished" podID="858b928e-a0c6-4c2d-9000-9e9807b7aecc" containerID="3295b9ab6a6ac3212d6e2f72c2016f956efa124402b317b46d6058bcf548f20e" exitCode=0 Sep 30 14:40:33 crc kubenswrapper[4840]: I0930 14:40:33.271245 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5jgck" event={"ID":"858b928e-a0c6-4c2d-9000-9e9807b7aecc","Type":"ContainerDied","Data":"3295b9ab6a6ac3212d6e2f72c2016f956efa124402b317b46d6058bcf548f20e"} Sep 30 14:40:33 crc kubenswrapper[4840]: I0930 14:40:33.271276 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5jgck" event={"ID":"858b928e-a0c6-4c2d-9000-9e9807b7aecc","Type":"ContainerDied","Data":"6db13533a63485c6755720f089cb43bcca34f27cc58a5d24a16c217c2041b8db"} Sep 30 14:40:33 crc kubenswrapper[4840]: I0930 14:40:33.271259 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5jgck" Sep 30 14:40:33 crc kubenswrapper[4840]: I0930 14:40:33.271296 4840 scope.go:117] "RemoveContainer" containerID="3295b9ab6a6ac3212d6e2f72c2016f956efa124402b317b46d6058bcf548f20e" Sep 30 14:40:33 crc kubenswrapper[4840]: I0930 14:40:33.348815 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/858b928e-a0c6-4c2d-9000-9e9807b7aecc-catalog-content\") pod \"858b928e-a0c6-4c2d-9000-9e9807b7aecc\" (UID: \"858b928e-a0c6-4c2d-9000-9e9807b7aecc\") " Sep 30 14:40:33 crc kubenswrapper[4840]: I0930 14:40:33.348901 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/858b928e-a0c6-4c2d-9000-9e9807b7aecc-utilities\") pod \"858b928e-a0c6-4c2d-9000-9e9807b7aecc\" (UID: \"858b928e-a0c6-4c2d-9000-9e9807b7aecc\") " Sep 30 14:40:33 crc kubenswrapper[4840]: I0930 14:40:33.349013 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8s6ss\" (UniqueName: \"kubernetes.io/projected/858b928e-a0c6-4c2d-9000-9e9807b7aecc-kube-api-access-8s6ss\") pod \"858b928e-a0c6-4c2d-9000-9e9807b7aecc\" (UID: \"858b928e-a0c6-4c2d-9000-9e9807b7aecc\") " Sep 30 14:40:33 crc kubenswrapper[4840]: I0930 14:40:33.349816 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/858b928e-a0c6-4c2d-9000-9e9807b7aecc-utilities" (OuterVolumeSpecName: "utilities") pod "858b928e-a0c6-4c2d-9000-9e9807b7aecc" (UID: "858b928e-a0c6-4c2d-9000-9e9807b7aecc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:40:33 crc kubenswrapper[4840]: I0930 14:40:33.354357 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/858b928e-a0c6-4c2d-9000-9e9807b7aecc-kube-api-access-8s6ss" (OuterVolumeSpecName: "kube-api-access-8s6ss") pod "858b928e-a0c6-4c2d-9000-9e9807b7aecc" (UID: "858b928e-a0c6-4c2d-9000-9e9807b7aecc"). InnerVolumeSpecName "kube-api-access-8s6ss". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:40:33 crc kubenswrapper[4840]: I0930 14:40:33.390697 4840 scope.go:117] "RemoveContainer" containerID="4e5391b00e315037cf2ccc869d34d297c823ca74294773b5a1b842141218b6a8" Sep 30 14:40:33 crc kubenswrapper[4840]: I0930 14:40:33.395732 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/858b928e-a0c6-4c2d-9000-9e9807b7aecc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "858b928e-a0c6-4c2d-9000-9e9807b7aecc" (UID: "858b928e-a0c6-4c2d-9000-9e9807b7aecc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:40:33 crc kubenswrapper[4840]: I0930 14:40:33.445499 4840 scope.go:117] "RemoveContainer" containerID="a27c73cb26c8132a1305dccdf2c5e9543b7f039e17389cf1a5d67a25dac35ab5" Sep 30 14:40:33 crc kubenswrapper[4840]: I0930 14:40:33.452008 4840 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/858b928e-a0c6-4c2d-9000-9e9807b7aecc-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:33 crc kubenswrapper[4840]: I0930 14:40:33.452038 4840 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/858b928e-a0c6-4c2d-9000-9e9807b7aecc-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:33 crc kubenswrapper[4840]: I0930 14:40:33.452048 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8s6ss\" (UniqueName: \"kubernetes.io/projected/858b928e-a0c6-4c2d-9000-9e9807b7aecc-kube-api-access-8s6ss\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:33 crc kubenswrapper[4840]: I0930 14:40:33.466732 4840 scope.go:117] "RemoveContainer" containerID="3295b9ab6a6ac3212d6e2f72c2016f956efa124402b317b46d6058bcf548f20e" Sep 30 14:40:33 crc kubenswrapper[4840]: E0930 14:40:33.467443 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3295b9ab6a6ac3212d6e2f72c2016f956efa124402b317b46d6058bcf548f20e\": container with ID starting with 3295b9ab6a6ac3212d6e2f72c2016f956efa124402b317b46d6058bcf548f20e not found: ID does not exist" containerID="3295b9ab6a6ac3212d6e2f72c2016f956efa124402b317b46d6058bcf548f20e" Sep 30 14:40:33 crc kubenswrapper[4840]: I0930 14:40:33.467534 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3295b9ab6a6ac3212d6e2f72c2016f956efa124402b317b46d6058bcf548f20e"} err="failed to get container status \"3295b9ab6a6ac3212d6e2f72c2016f956efa124402b317b46d6058bcf548f20e\": rpc error: code = NotFound desc = could not find container \"3295b9ab6a6ac3212d6e2f72c2016f956efa124402b317b46d6058bcf548f20e\": container with ID starting with 3295b9ab6a6ac3212d6e2f72c2016f956efa124402b317b46d6058bcf548f20e not found: ID does not exist" Sep 30 14:40:33 crc kubenswrapper[4840]: I0930 14:40:33.467618 4840 scope.go:117] "RemoveContainer" containerID="4e5391b00e315037cf2ccc869d34d297c823ca74294773b5a1b842141218b6a8" Sep 30 14:40:33 crc kubenswrapper[4840]: E0930 14:40:33.468070 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e5391b00e315037cf2ccc869d34d297c823ca74294773b5a1b842141218b6a8\": container with ID starting with 4e5391b00e315037cf2ccc869d34d297c823ca74294773b5a1b842141218b6a8 not found: ID does not exist" containerID="4e5391b00e315037cf2ccc869d34d297c823ca74294773b5a1b842141218b6a8" Sep 30 14:40:33 crc kubenswrapper[4840]: I0930 14:40:33.468141 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e5391b00e315037cf2ccc869d34d297c823ca74294773b5a1b842141218b6a8"} err="failed to get container status \"4e5391b00e315037cf2ccc869d34d297c823ca74294773b5a1b842141218b6a8\": rpc error: code = NotFound desc = could not find container \"4e5391b00e315037cf2ccc869d34d297c823ca74294773b5a1b842141218b6a8\": container with ID starting with 4e5391b00e315037cf2ccc869d34d297c823ca74294773b5a1b842141218b6a8 not found: ID does not exist" Sep 30 14:40:33 crc kubenswrapper[4840]: I0930 14:40:33.468199 4840 scope.go:117] "RemoveContainer" containerID="a27c73cb26c8132a1305dccdf2c5e9543b7f039e17389cf1a5d67a25dac35ab5" Sep 30 14:40:33 crc kubenswrapper[4840]: E0930 14:40:33.468668 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a27c73cb26c8132a1305dccdf2c5e9543b7f039e17389cf1a5d67a25dac35ab5\": container with ID starting with a27c73cb26c8132a1305dccdf2c5e9543b7f039e17389cf1a5d67a25dac35ab5 not found: ID does not exist" containerID="a27c73cb26c8132a1305dccdf2c5e9543b7f039e17389cf1a5d67a25dac35ab5" Sep 30 14:40:33 crc kubenswrapper[4840]: I0930 14:40:33.468717 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a27c73cb26c8132a1305dccdf2c5e9543b7f039e17389cf1a5d67a25dac35ab5"} err="failed to get container status \"a27c73cb26c8132a1305dccdf2c5e9543b7f039e17389cf1a5d67a25dac35ab5\": rpc error: code = NotFound desc = could not find container \"a27c73cb26c8132a1305dccdf2c5e9543b7f039e17389cf1a5d67a25dac35ab5\": container with ID starting with a27c73cb26c8132a1305dccdf2c5e9543b7f039e17389cf1a5d67a25dac35ab5 not found: ID does not exist" Sep 30 14:40:33 crc kubenswrapper[4840]: I0930 14:40:33.604643 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5jgck"] Sep 30 14:40:33 crc kubenswrapper[4840]: I0930 14:40:33.614392 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5jgck"] Sep 30 14:40:34 crc kubenswrapper[4840]: I0930 14:40:34.126901 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="858b928e-a0c6-4c2d-9000-9e9807b7aecc" path="/var/lib/kubelet/pods/858b928e-a0c6-4c2d-9000-9e9807b7aecc/volumes" Sep 30 14:40:39 crc kubenswrapper[4840]: I0930 14:40:39.116565 4840 scope.go:117] "RemoveContainer" containerID="e1e2bfc7f9537971d768f1b7183216797522532290c5822edafa8e0b5232aa98" Sep 30 14:40:39 crc kubenswrapper[4840]: E0930 14:40:39.117386 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:40:51 crc kubenswrapper[4840]: E0930 14:40:51.592827 4840 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Sep 30 14:40:51 crc kubenswrapper[4840]: E0930 14:40:51.593605 4840 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hnfwb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 14:40:51 crc kubenswrapper[4840]: E0930 14:40:51.594958 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545" Sep 30 14:40:52 crc kubenswrapper[4840]: E0930 14:40:52.455517 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545" Sep 30 14:40:53 crc kubenswrapper[4840]: I0930 14:40:53.116144 4840 scope.go:117] "RemoveContainer" containerID="e1e2bfc7f9537971d768f1b7183216797522532290c5822edafa8e0b5232aa98" Sep 30 14:40:53 crc kubenswrapper[4840]: E0930 14:40:53.116528 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:41:04 crc kubenswrapper[4840]: I0930 14:41:04.116535 4840 scope.go:117] "RemoveContainer" containerID="e1e2bfc7f9537971d768f1b7183216797522532290c5822edafa8e0b5232aa98" Sep 30 14:41:04 crc kubenswrapper[4840]: E0930 14:41:04.117646 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:41:06 crc kubenswrapper[4840]: I0930 14:41:06.570484 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545","Type":"ContainerStarted","Data":"974d10e15595c7d52e8b9180969026db815eba4e38f1c22f60f533ae5e0f056a"} Sep 30 14:41:06 crc kubenswrapper[4840]: I0930 14:41:06.586942 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.369227963 podStartE2EDuration="44.586926569s" podCreationTimestamp="2025-09-30 14:40:22 +0000 UTC" firstStartedPulling="2025-09-30 14:40:24.438692248 +0000 UTC m=+2653.067778691" lastFinishedPulling="2025-09-30 14:41:05.656390874 +0000 UTC m=+2694.285477297" observedRunningTime="2025-09-30 14:41:06.584631424 +0000 UTC m=+2695.213717857" watchObservedRunningTime="2025-09-30 14:41:06.586926569 +0000 UTC m=+2695.216012992" Sep 30 14:41:19 crc kubenswrapper[4840]: I0930 14:41:19.116923 4840 scope.go:117] "RemoveContainer" containerID="e1e2bfc7f9537971d768f1b7183216797522532290c5822edafa8e0b5232aa98" Sep 30 14:41:19 crc kubenswrapper[4840]: E0930 14:41:19.117867 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:41:32 crc kubenswrapper[4840]: I0930 14:41:32.125674 4840 scope.go:117] "RemoveContainer" containerID="e1e2bfc7f9537971d768f1b7183216797522532290c5822edafa8e0b5232aa98" Sep 30 14:41:32 crc kubenswrapper[4840]: I0930 14:41:32.799626 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" event={"ID":"10e8b890-7f20-4a36-8e03-898620cf599a","Type":"ContainerStarted","Data":"114f6accf72be8011201b200dc40143e1ba0f13c3f60c2a725c667df5ceff985"} Sep 30 14:43:30 crc kubenswrapper[4840]: I0930 14:43:30.970835 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lv45c"] Sep 30 14:43:30 crc kubenswrapper[4840]: E0930 14:43:30.971847 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="858b928e-a0c6-4c2d-9000-9e9807b7aecc" containerName="extract-utilities" Sep 30 14:43:30 crc kubenswrapper[4840]: I0930 14:43:30.971861 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="858b928e-a0c6-4c2d-9000-9e9807b7aecc" containerName="extract-utilities" Sep 30 14:43:30 crc kubenswrapper[4840]: E0930 14:43:30.971877 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="858b928e-a0c6-4c2d-9000-9e9807b7aecc" containerName="extract-content" Sep 30 14:43:30 crc kubenswrapper[4840]: I0930 14:43:30.971883 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="858b928e-a0c6-4c2d-9000-9e9807b7aecc" containerName="extract-content" Sep 30 14:43:30 crc kubenswrapper[4840]: E0930 14:43:30.971913 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="858b928e-a0c6-4c2d-9000-9e9807b7aecc" containerName="registry-server" Sep 30 14:43:30 crc kubenswrapper[4840]: I0930 14:43:30.971919 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="858b928e-a0c6-4c2d-9000-9e9807b7aecc" containerName="registry-server" Sep 30 14:43:30 crc kubenswrapper[4840]: I0930 14:43:30.972107 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="858b928e-a0c6-4c2d-9000-9e9807b7aecc" containerName="registry-server" Sep 30 14:43:30 crc kubenswrapper[4840]: I0930 14:43:30.973774 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lv45c" Sep 30 14:43:30 crc kubenswrapper[4840]: I0930 14:43:30.986990 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lv45c"] Sep 30 14:43:31 crc kubenswrapper[4840]: I0930 14:43:31.172472 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5p5g\" (UniqueName: \"kubernetes.io/projected/a0f32d36-c239-40e6-bfeb-f53c50391beb-kube-api-access-n5p5g\") pod \"community-operators-lv45c\" (UID: \"a0f32d36-c239-40e6-bfeb-f53c50391beb\") " pod="openshift-marketplace/community-operators-lv45c" Sep 30 14:43:31 crc kubenswrapper[4840]: I0930 14:43:31.173421 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0f32d36-c239-40e6-bfeb-f53c50391beb-catalog-content\") pod \"community-operators-lv45c\" (UID: \"a0f32d36-c239-40e6-bfeb-f53c50391beb\") " pod="openshift-marketplace/community-operators-lv45c" Sep 30 14:43:31 crc kubenswrapper[4840]: I0930 14:43:31.173749 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0f32d36-c239-40e6-bfeb-f53c50391beb-utilities\") pod \"community-operators-lv45c\" (UID: \"a0f32d36-c239-40e6-bfeb-f53c50391beb\") " pod="openshift-marketplace/community-operators-lv45c" Sep 30 14:43:31 crc kubenswrapper[4840]: I0930 14:43:31.285939 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5p5g\" (UniqueName: \"kubernetes.io/projected/a0f32d36-c239-40e6-bfeb-f53c50391beb-kube-api-access-n5p5g\") pod \"community-operators-lv45c\" (UID: \"a0f32d36-c239-40e6-bfeb-f53c50391beb\") " pod="openshift-marketplace/community-operators-lv45c" Sep 30 14:43:31 crc kubenswrapper[4840]: I0930 14:43:31.286319 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0f32d36-c239-40e6-bfeb-f53c50391beb-catalog-content\") pod \"community-operators-lv45c\" (UID: \"a0f32d36-c239-40e6-bfeb-f53c50391beb\") " pod="openshift-marketplace/community-operators-lv45c" Sep 30 14:43:31 crc kubenswrapper[4840]: I0930 14:43:31.286358 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0f32d36-c239-40e6-bfeb-f53c50391beb-utilities\") pod \"community-operators-lv45c\" (UID: \"a0f32d36-c239-40e6-bfeb-f53c50391beb\") " pod="openshift-marketplace/community-operators-lv45c" Sep 30 14:43:31 crc kubenswrapper[4840]: I0930 14:43:31.287128 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0f32d36-c239-40e6-bfeb-f53c50391beb-utilities\") pod \"community-operators-lv45c\" (UID: \"a0f32d36-c239-40e6-bfeb-f53c50391beb\") " pod="openshift-marketplace/community-operators-lv45c" Sep 30 14:43:31 crc kubenswrapper[4840]: I0930 14:43:31.288950 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0f32d36-c239-40e6-bfeb-f53c50391beb-catalog-content\") pod \"community-operators-lv45c\" (UID: \"a0f32d36-c239-40e6-bfeb-f53c50391beb\") " pod="openshift-marketplace/community-operators-lv45c" Sep 30 14:43:31 crc kubenswrapper[4840]: I0930 14:43:31.341378 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5p5g\" (UniqueName: \"kubernetes.io/projected/a0f32d36-c239-40e6-bfeb-f53c50391beb-kube-api-access-n5p5g\") pod \"community-operators-lv45c\" (UID: \"a0f32d36-c239-40e6-bfeb-f53c50391beb\") " pod="openshift-marketplace/community-operators-lv45c" Sep 30 14:43:31 crc kubenswrapper[4840]: I0930 14:43:31.597926 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lv45c" Sep 30 14:43:32 crc kubenswrapper[4840]: I0930 14:43:32.072542 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lv45c"] Sep 30 14:43:32 crc kubenswrapper[4840]: I0930 14:43:32.867233 4840 generic.go:334] "Generic (PLEG): container finished" podID="a0f32d36-c239-40e6-bfeb-f53c50391beb" containerID="4866d791a0f8a49b40577097349fbde2418539af2e5a4c2d8a4a7c8843a0c0d7" exitCode=0 Sep 30 14:43:32 crc kubenswrapper[4840]: I0930 14:43:32.867324 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lv45c" event={"ID":"a0f32d36-c239-40e6-bfeb-f53c50391beb","Type":"ContainerDied","Data":"4866d791a0f8a49b40577097349fbde2418539af2e5a4c2d8a4a7c8843a0c0d7"} Sep 30 14:43:32 crc kubenswrapper[4840]: I0930 14:43:32.867664 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lv45c" event={"ID":"a0f32d36-c239-40e6-bfeb-f53c50391beb","Type":"ContainerStarted","Data":"09afaad98390aaa15d75a624ae60d79f7f8dd4779482f29eb5dab00084e06c13"} Sep 30 14:43:34 crc kubenswrapper[4840]: I0930 14:43:34.888130 4840 generic.go:334] "Generic (PLEG): container finished" podID="a0f32d36-c239-40e6-bfeb-f53c50391beb" containerID="0aeefc8166b123e496626ce0efd382d973d2fec43fdc314a2862a233ef88016f" exitCode=0 Sep 30 14:43:34 crc kubenswrapper[4840]: I0930 14:43:34.888392 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lv45c" event={"ID":"a0f32d36-c239-40e6-bfeb-f53c50391beb","Type":"ContainerDied","Data":"0aeefc8166b123e496626ce0efd382d973d2fec43fdc314a2862a233ef88016f"} Sep 30 14:43:35 crc kubenswrapper[4840]: I0930 14:43:35.902754 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lv45c" event={"ID":"a0f32d36-c239-40e6-bfeb-f53c50391beb","Type":"ContainerStarted","Data":"1538128b6bdfe90baf9e7e4c91204263198bd70c4475558c5db67a4757cc2fa7"} Sep 30 14:43:35 crc kubenswrapper[4840]: I0930 14:43:35.935216 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lv45c" podStartSLOduration=3.338644902 podStartE2EDuration="5.93519395s" podCreationTimestamp="2025-09-30 14:43:30 +0000 UTC" firstStartedPulling="2025-09-30 14:43:32.868833117 +0000 UTC m=+2841.497919540" lastFinishedPulling="2025-09-30 14:43:35.465382165 +0000 UTC m=+2844.094468588" observedRunningTime="2025-09-30 14:43:35.926389435 +0000 UTC m=+2844.555475858" watchObservedRunningTime="2025-09-30 14:43:35.93519395 +0000 UTC m=+2844.564280373" Sep 30 14:43:41 crc kubenswrapper[4840]: I0930 14:43:41.598590 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lv45c" Sep 30 14:43:41 crc kubenswrapper[4840]: I0930 14:43:41.600347 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lv45c" Sep 30 14:43:41 crc kubenswrapper[4840]: I0930 14:43:41.649460 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lv45c" Sep 30 14:43:41 crc kubenswrapper[4840]: I0930 14:43:41.995010 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lv45c" Sep 30 14:43:42 crc kubenswrapper[4840]: I0930 14:43:42.047371 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lv45c"] Sep 30 14:43:43 crc kubenswrapper[4840]: I0930 14:43:43.967918 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lv45c" podUID="a0f32d36-c239-40e6-bfeb-f53c50391beb" containerName="registry-server" containerID="cri-o://1538128b6bdfe90baf9e7e4c91204263198bd70c4475558c5db67a4757cc2fa7" gracePeriod=2 Sep 30 14:43:44 crc kubenswrapper[4840]: I0930 14:43:44.465504 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lv45c" Sep 30 14:43:44 crc kubenswrapper[4840]: I0930 14:43:44.644866 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5p5g\" (UniqueName: \"kubernetes.io/projected/a0f32d36-c239-40e6-bfeb-f53c50391beb-kube-api-access-n5p5g\") pod \"a0f32d36-c239-40e6-bfeb-f53c50391beb\" (UID: \"a0f32d36-c239-40e6-bfeb-f53c50391beb\") " Sep 30 14:43:44 crc kubenswrapper[4840]: I0930 14:43:44.644991 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0f32d36-c239-40e6-bfeb-f53c50391beb-utilities\") pod \"a0f32d36-c239-40e6-bfeb-f53c50391beb\" (UID: \"a0f32d36-c239-40e6-bfeb-f53c50391beb\") " Sep 30 14:43:44 crc kubenswrapper[4840]: I0930 14:43:44.645018 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0f32d36-c239-40e6-bfeb-f53c50391beb-catalog-content\") pod \"a0f32d36-c239-40e6-bfeb-f53c50391beb\" (UID: \"a0f32d36-c239-40e6-bfeb-f53c50391beb\") " Sep 30 14:43:44 crc kubenswrapper[4840]: I0930 14:43:44.645934 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0f32d36-c239-40e6-bfeb-f53c50391beb-utilities" (OuterVolumeSpecName: "utilities") pod "a0f32d36-c239-40e6-bfeb-f53c50391beb" (UID: "a0f32d36-c239-40e6-bfeb-f53c50391beb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:43:44 crc kubenswrapper[4840]: I0930 14:43:44.651037 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0f32d36-c239-40e6-bfeb-f53c50391beb-kube-api-access-n5p5g" (OuterVolumeSpecName: "kube-api-access-n5p5g") pod "a0f32d36-c239-40e6-bfeb-f53c50391beb" (UID: "a0f32d36-c239-40e6-bfeb-f53c50391beb"). InnerVolumeSpecName "kube-api-access-n5p5g". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:43:44 crc kubenswrapper[4840]: I0930 14:43:44.698524 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0f32d36-c239-40e6-bfeb-f53c50391beb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a0f32d36-c239-40e6-bfeb-f53c50391beb" (UID: "a0f32d36-c239-40e6-bfeb-f53c50391beb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:43:44 crc kubenswrapper[4840]: I0930 14:43:44.747734 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5p5g\" (UniqueName: \"kubernetes.io/projected/a0f32d36-c239-40e6-bfeb-f53c50391beb-kube-api-access-n5p5g\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:44 crc kubenswrapper[4840]: I0930 14:43:44.747780 4840 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0f32d36-c239-40e6-bfeb-f53c50391beb-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:44 crc kubenswrapper[4840]: I0930 14:43:44.747792 4840 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0f32d36-c239-40e6-bfeb-f53c50391beb-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:44 crc kubenswrapper[4840]: I0930 14:43:44.978954 4840 generic.go:334] "Generic (PLEG): container finished" podID="a0f32d36-c239-40e6-bfeb-f53c50391beb" containerID="1538128b6bdfe90baf9e7e4c91204263198bd70c4475558c5db67a4757cc2fa7" exitCode=0 Sep 30 14:43:44 crc kubenswrapper[4840]: I0930 14:43:44.978999 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lv45c" event={"ID":"a0f32d36-c239-40e6-bfeb-f53c50391beb","Type":"ContainerDied","Data":"1538128b6bdfe90baf9e7e4c91204263198bd70c4475558c5db67a4757cc2fa7"} Sep 30 14:43:44 crc kubenswrapper[4840]: I0930 14:43:44.979025 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lv45c" event={"ID":"a0f32d36-c239-40e6-bfeb-f53c50391beb","Type":"ContainerDied","Data":"09afaad98390aaa15d75a624ae60d79f7f8dd4779482f29eb5dab00084e06c13"} Sep 30 14:43:44 crc kubenswrapper[4840]: I0930 14:43:44.979024 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lv45c" Sep 30 14:43:44 crc kubenswrapper[4840]: I0930 14:43:44.979041 4840 scope.go:117] "RemoveContainer" containerID="1538128b6bdfe90baf9e7e4c91204263198bd70c4475558c5db67a4757cc2fa7" Sep 30 14:43:45 crc kubenswrapper[4840]: I0930 14:43:45.004865 4840 scope.go:117] "RemoveContainer" containerID="0aeefc8166b123e496626ce0efd382d973d2fec43fdc314a2862a233ef88016f" Sep 30 14:43:45 crc kubenswrapper[4840]: I0930 14:43:45.020507 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lv45c"] Sep 30 14:43:45 crc kubenswrapper[4840]: I0930 14:43:45.029604 4840 scope.go:117] "RemoveContainer" containerID="4866d791a0f8a49b40577097349fbde2418539af2e5a4c2d8a4a7c8843a0c0d7" Sep 30 14:43:45 crc kubenswrapper[4840]: I0930 14:43:45.029764 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lv45c"] Sep 30 14:43:45 crc kubenswrapper[4840]: I0930 14:43:45.073787 4840 scope.go:117] "RemoveContainer" containerID="1538128b6bdfe90baf9e7e4c91204263198bd70c4475558c5db67a4757cc2fa7" Sep 30 14:43:45 crc kubenswrapper[4840]: E0930 14:43:45.074332 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1538128b6bdfe90baf9e7e4c91204263198bd70c4475558c5db67a4757cc2fa7\": container with ID starting with 1538128b6bdfe90baf9e7e4c91204263198bd70c4475558c5db67a4757cc2fa7 not found: ID does not exist" containerID="1538128b6bdfe90baf9e7e4c91204263198bd70c4475558c5db67a4757cc2fa7" Sep 30 14:43:45 crc kubenswrapper[4840]: I0930 14:43:45.074378 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1538128b6bdfe90baf9e7e4c91204263198bd70c4475558c5db67a4757cc2fa7"} err="failed to get container status \"1538128b6bdfe90baf9e7e4c91204263198bd70c4475558c5db67a4757cc2fa7\": rpc error: code = NotFound desc = could not find container \"1538128b6bdfe90baf9e7e4c91204263198bd70c4475558c5db67a4757cc2fa7\": container with ID starting with 1538128b6bdfe90baf9e7e4c91204263198bd70c4475558c5db67a4757cc2fa7 not found: ID does not exist" Sep 30 14:43:45 crc kubenswrapper[4840]: I0930 14:43:45.074413 4840 scope.go:117] "RemoveContainer" containerID="0aeefc8166b123e496626ce0efd382d973d2fec43fdc314a2862a233ef88016f" Sep 30 14:43:45 crc kubenswrapper[4840]: E0930 14:43:45.075174 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0aeefc8166b123e496626ce0efd382d973d2fec43fdc314a2862a233ef88016f\": container with ID starting with 0aeefc8166b123e496626ce0efd382d973d2fec43fdc314a2862a233ef88016f not found: ID does not exist" containerID="0aeefc8166b123e496626ce0efd382d973d2fec43fdc314a2862a233ef88016f" Sep 30 14:43:45 crc kubenswrapper[4840]: I0930 14:43:45.075217 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0aeefc8166b123e496626ce0efd382d973d2fec43fdc314a2862a233ef88016f"} err="failed to get container status \"0aeefc8166b123e496626ce0efd382d973d2fec43fdc314a2862a233ef88016f\": rpc error: code = NotFound desc = could not find container \"0aeefc8166b123e496626ce0efd382d973d2fec43fdc314a2862a233ef88016f\": container with ID starting with 0aeefc8166b123e496626ce0efd382d973d2fec43fdc314a2862a233ef88016f not found: ID does not exist" Sep 30 14:43:45 crc kubenswrapper[4840]: I0930 14:43:45.075235 4840 scope.go:117] "RemoveContainer" containerID="4866d791a0f8a49b40577097349fbde2418539af2e5a4c2d8a4a7c8843a0c0d7" Sep 30 14:43:45 crc kubenswrapper[4840]: E0930 14:43:45.075502 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4866d791a0f8a49b40577097349fbde2418539af2e5a4c2d8a4a7c8843a0c0d7\": container with ID starting with 4866d791a0f8a49b40577097349fbde2418539af2e5a4c2d8a4a7c8843a0c0d7 not found: ID does not exist" containerID="4866d791a0f8a49b40577097349fbde2418539af2e5a4c2d8a4a7c8843a0c0d7" Sep 30 14:43:45 crc kubenswrapper[4840]: I0930 14:43:45.075522 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4866d791a0f8a49b40577097349fbde2418539af2e5a4c2d8a4a7c8843a0c0d7"} err="failed to get container status \"4866d791a0f8a49b40577097349fbde2418539af2e5a4c2d8a4a7c8843a0c0d7\": rpc error: code = NotFound desc = could not find container \"4866d791a0f8a49b40577097349fbde2418539af2e5a4c2d8a4a7c8843a0c0d7\": container with ID starting with 4866d791a0f8a49b40577097349fbde2418539af2e5a4c2d8a4a7c8843a0c0d7 not found: ID does not exist" Sep 30 14:43:46 crc kubenswrapper[4840]: I0930 14:43:46.128878 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0f32d36-c239-40e6-bfeb-f53c50391beb" path="/var/lib/kubelet/pods/a0f32d36-c239-40e6-bfeb-f53c50391beb/volumes" Sep 30 14:43:51 crc kubenswrapper[4840]: I0930 14:43:51.871484 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:43:51 crc kubenswrapper[4840]: I0930 14:43:51.872166 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:44:21 crc kubenswrapper[4840]: I0930 14:44:21.871647 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:44:21 crc kubenswrapper[4840]: I0930 14:44:21.872327 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:44:51 crc kubenswrapper[4840]: I0930 14:44:51.871635 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:44:51 crc kubenswrapper[4840]: I0930 14:44:51.872501 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:44:51 crc kubenswrapper[4840]: I0930 14:44:51.872592 4840 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-747gk" Sep 30 14:44:51 crc kubenswrapper[4840]: I0930 14:44:51.873602 4840 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"114f6accf72be8011201b200dc40143e1ba0f13c3f60c2a725c667df5ceff985"} pod="openshift-machine-config-operator/machine-config-daemon-747gk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 14:44:51 crc kubenswrapper[4840]: I0930 14:44:51.873704 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" containerID="cri-o://114f6accf72be8011201b200dc40143e1ba0f13c3f60c2a725c667df5ceff985" gracePeriod=600 Sep 30 14:44:52 crc kubenswrapper[4840]: I0930 14:44:52.574968 4840 generic.go:334] "Generic (PLEG): container finished" podID="10e8b890-7f20-4a36-8e03-898620cf599a" containerID="114f6accf72be8011201b200dc40143e1ba0f13c3f60c2a725c667df5ceff985" exitCode=0 Sep 30 14:44:52 crc kubenswrapper[4840]: I0930 14:44:52.575047 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" event={"ID":"10e8b890-7f20-4a36-8e03-898620cf599a","Type":"ContainerDied","Data":"114f6accf72be8011201b200dc40143e1ba0f13c3f60c2a725c667df5ceff985"} Sep 30 14:44:52 crc kubenswrapper[4840]: I0930 14:44:52.575377 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" event={"ID":"10e8b890-7f20-4a36-8e03-898620cf599a","Type":"ContainerStarted","Data":"bb42e5820a8aaddbc79e0404192d904aa52fd67a0816b292b3605d7cf3a9d5b0"} Sep 30 14:44:52 crc kubenswrapper[4840]: I0930 14:44:52.575403 4840 scope.go:117] "RemoveContainer" containerID="e1e2bfc7f9537971d768f1b7183216797522532290c5822edafa8e0b5232aa98" Sep 30 14:45:00 crc kubenswrapper[4840]: I0930 14:45:00.192886 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320725-c6lmj"] Sep 30 14:45:00 crc kubenswrapper[4840]: E0930 14:45:00.194094 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0f32d36-c239-40e6-bfeb-f53c50391beb" containerName="extract-utilities" Sep 30 14:45:00 crc kubenswrapper[4840]: I0930 14:45:00.194113 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0f32d36-c239-40e6-bfeb-f53c50391beb" containerName="extract-utilities" Sep 30 14:45:00 crc kubenswrapper[4840]: E0930 14:45:00.194134 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0f32d36-c239-40e6-bfeb-f53c50391beb" containerName="extract-content" Sep 30 14:45:00 crc kubenswrapper[4840]: I0930 14:45:00.194142 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0f32d36-c239-40e6-bfeb-f53c50391beb" containerName="extract-content" Sep 30 14:45:00 crc kubenswrapper[4840]: E0930 14:45:00.194184 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0f32d36-c239-40e6-bfeb-f53c50391beb" containerName="registry-server" Sep 30 14:45:00 crc kubenswrapper[4840]: I0930 14:45:00.194195 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0f32d36-c239-40e6-bfeb-f53c50391beb" containerName="registry-server" Sep 30 14:45:00 crc kubenswrapper[4840]: I0930 14:45:00.194391 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0f32d36-c239-40e6-bfeb-f53c50391beb" containerName="registry-server" Sep 30 14:45:00 crc kubenswrapper[4840]: I0930 14:45:00.195200 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320725-c6lmj" Sep 30 14:45:00 crc kubenswrapper[4840]: I0930 14:45:00.197417 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 30 14:45:00 crc kubenswrapper[4840]: I0930 14:45:00.201058 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 30 14:45:00 crc kubenswrapper[4840]: I0930 14:45:00.201817 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320725-c6lmj"] Sep 30 14:45:00 crc kubenswrapper[4840]: I0930 14:45:00.353902 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6knx\" (UniqueName: \"kubernetes.io/projected/1e1e0905-68d7-4038-91fa-3c9934c5cccc-kube-api-access-r6knx\") pod \"collect-profiles-29320725-c6lmj\" (UID: \"1e1e0905-68d7-4038-91fa-3c9934c5cccc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320725-c6lmj" Sep 30 14:45:00 crc kubenswrapper[4840]: I0930 14:45:00.354716 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1e1e0905-68d7-4038-91fa-3c9934c5cccc-secret-volume\") pod \"collect-profiles-29320725-c6lmj\" (UID: \"1e1e0905-68d7-4038-91fa-3c9934c5cccc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320725-c6lmj" Sep 30 14:45:00 crc kubenswrapper[4840]: I0930 14:45:00.354971 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e1e0905-68d7-4038-91fa-3c9934c5cccc-config-volume\") pod \"collect-profiles-29320725-c6lmj\" (UID: \"1e1e0905-68d7-4038-91fa-3c9934c5cccc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320725-c6lmj" Sep 30 14:45:00 crc kubenswrapper[4840]: I0930 14:45:00.457226 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e1e0905-68d7-4038-91fa-3c9934c5cccc-config-volume\") pod \"collect-profiles-29320725-c6lmj\" (UID: \"1e1e0905-68d7-4038-91fa-3c9934c5cccc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320725-c6lmj" Sep 30 14:45:00 crc kubenswrapper[4840]: I0930 14:45:00.457395 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6knx\" (UniqueName: \"kubernetes.io/projected/1e1e0905-68d7-4038-91fa-3c9934c5cccc-kube-api-access-r6knx\") pod \"collect-profiles-29320725-c6lmj\" (UID: \"1e1e0905-68d7-4038-91fa-3c9934c5cccc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320725-c6lmj" Sep 30 14:45:00 crc kubenswrapper[4840]: I0930 14:45:00.457421 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1e1e0905-68d7-4038-91fa-3c9934c5cccc-secret-volume\") pod \"collect-profiles-29320725-c6lmj\" (UID: \"1e1e0905-68d7-4038-91fa-3c9934c5cccc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320725-c6lmj" Sep 30 14:45:00 crc kubenswrapper[4840]: I0930 14:45:00.459376 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e1e0905-68d7-4038-91fa-3c9934c5cccc-config-volume\") pod \"collect-profiles-29320725-c6lmj\" (UID: \"1e1e0905-68d7-4038-91fa-3c9934c5cccc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320725-c6lmj" Sep 30 14:45:00 crc kubenswrapper[4840]: I0930 14:45:00.465687 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1e1e0905-68d7-4038-91fa-3c9934c5cccc-secret-volume\") pod \"collect-profiles-29320725-c6lmj\" (UID: \"1e1e0905-68d7-4038-91fa-3c9934c5cccc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320725-c6lmj" Sep 30 14:45:00 crc kubenswrapper[4840]: I0930 14:45:00.475641 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6knx\" (UniqueName: \"kubernetes.io/projected/1e1e0905-68d7-4038-91fa-3c9934c5cccc-kube-api-access-r6knx\") pod \"collect-profiles-29320725-c6lmj\" (UID: \"1e1e0905-68d7-4038-91fa-3c9934c5cccc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320725-c6lmj" Sep 30 14:45:00 crc kubenswrapper[4840]: I0930 14:45:00.530008 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320725-c6lmj" Sep 30 14:45:00 crc kubenswrapper[4840]: I0930 14:45:00.974664 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320725-c6lmj"] Sep 30 14:45:01 crc kubenswrapper[4840]: I0930 14:45:01.667376 4840 generic.go:334] "Generic (PLEG): container finished" podID="1e1e0905-68d7-4038-91fa-3c9934c5cccc" containerID="2238466b3043d026ab0b2f8e96b42dc509fdb8fde5ac4b418bf67ba886fde59c" exitCode=0 Sep 30 14:45:01 crc kubenswrapper[4840]: I0930 14:45:01.667465 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320725-c6lmj" event={"ID":"1e1e0905-68d7-4038-91fa-3c9934c5cccc","Type":"ContainerDied","Data":"2238466b3043d026ab0b2f8e96b42dc509fdb8fde5ac4b418bf67ba886fde59c"} Sep 30 14:45:01 crc kubenswrapper[4840]: I0930 14:45:01.667984 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320725-c6lmj" event={"ID":"1e1e0905-68d7-4038-91fa-3c9934c5cccc","Type":"ContainerStarted","Data":"1445a2c54c8d2dd4420e0773c4721923c8ef8f3cb6a99e70e4b6b84de6565254"} Sep 30 14:45:03 crc kubenswrapper[4840]: I0930 14:45:03.080842 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320725-c6lmj" Sep 30 14:45:03 crc kubenswrapper[4840]: I0930 14:45:03.206963 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6knx\" (UniqueName: \"kubernetes.io/projected/1e1e0905-68d7-4038-91fa-3c9934c5cccc-kube-api-access-r6knx\") pod \"1e1e0905-68d7-4038-91fa-3c9934c5cccc\" (UID: \"1e1e0905-68d7-4038-91fa-3c9934c5cccc\") " Sep 30 14:45:03 crc kubenswrapper[4840]: I0930 14:45:03.207068 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e1e0905-68d7-4038-91fa-3c9934c5cccc-config-volume\") pod \"1e1e0905-68d7-4038-91fa-3c9934c5cccc\" (UID: \"1e1e0905-68d7-4038-91fa-3c9934c5cccc\") " Sep 30 14:45:03 crc kubenswrapper[4840]: I0930 14:45:03.207199 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1e1e0905-68d7-4038-91fa-3c9934c5cccc-secret-volume\") pod \"1e1e0905-68d7-4038-91fa-3c9934c5cccc\" (UID: \"1e1e0905-68d7-4038-91fa-3c9934c5cccc\") " Sep 30 14:45:03 crc kubenswrapper[4840]: I0930 14:45:03.207903 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e1e0905-68d7-4038-91fa-3c9934c5cccc-config-volume" (OuterVolumeSpecName: "config-volume") pod "1e1e0905-68d7-4038-91fa-3c9934c5cccc" (UID: "1e1e0905-68d7-4038-91fa-3c9934c5cccc"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:45:03 crc kubenswrapper[4840]: I0930 14:45:03.213396 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e1e0905-68d7-4038-91fa-3c9934c5cccc-kube-api-access-r6knx" (OuterVolumeSpecName: "kube-api-access-r6knx") pod "1e1e0905-68d7-4038-91fa-3c9934c5cccc" (UID: "1e1e0905-68d7-4038-91fa-3c9934c5cccc"). InnerVolumeSpecName "kube-api-access-r6knx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:45:03 crc kubenswrapper[4840]: I0930 14:45:03.213781 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e1e0905-68d7-4038-91fa-3c9934c5cccc-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1e1e0905-68d7-4038-91fa-3c9934c5cccc" (UID: "1e1e0905-68d7-4038-91fa-3c9934c5cccc"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:45:03 crc kubenswrapper[4840]: I0930 14:45:03.310008 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6knx\" (UniqueName: \"kubernetes.io/projected/1e1e0905-68d7-4038-91fa-3c9934c5cccc-kube-api-access-r6knx\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:03 crc kubenswrapper[4840]: I0930 14:45:03.310041 4840 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e1e0905-68d7-4038-91fa-3c9934c5cccc-config-volume\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:03 crc kubenswrapper[4840]: I0930 14:45:03.310050 4840 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1e1e0905-68d7-4038-91fa-3c9934c5cccc-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:03 crc kubenswrapper[4840]: I0930 14:45:03.684761 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320725-c6lmj" event={"ID":"1e1e0905-68d7-4038-91fa-3c9934c5cccc","Type":"ContainerDied","Data":"1445a2c54c8d2dd4420e0773c4721923c8ef8f3cb6a99e70e4b6b84de6565254"} Sep 30 14:45:03 crc kubenswrapper[4840]: I0930 14:45:03.684812 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320725-c6lmj" Sep 30 14:45:03 crc kubenswrapper[4840]: I0930 14:45:03.684827 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1445a2c54c8d2dd4420e0773c4721923c8ef8f3cb6a99e70e4b6b84de6565254" Sep 30 14:45:04 crc kubenswrapper[4840]: I0930 14:45:04.155949 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320680-vtwpz"] Sep 30 14:45:04 crc kubenswrapper[4840]: I0930 14:45:04.164289 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320680-vtwpz"] Sep 30 14:45:06 crc kubenswrapper[4840]: I0930 14:45:06.128709 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="556ac8c1-39c4-4a47-a542-6881f953ef8f" path="/var/lib/kubelet/pods/556ac8c1-39c4-4a47-a542-6881f953ef8f/volumes" Sep 30 14:45:51 crc kubenswrapper[4840]: I0930 14:45:51.622623 4840 scope.go:117] "RemoveContainer" containerID="87134cf6d6900bd5d20a83e4fe6e39464405a640d3334de5c538f352c5bd5dce" Sep 30 14:47:21 crc kubenswrapper[4840]: I0930 14:47:21.871483 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:47:21 crc kubenswrapper[4840]: I0930 14:47:21.872077 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:47:31 crc kubenswrapper[4840]: I0930 14:47:31.383872 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-f5hcj"] Sep 30 14:47:31 crc kubenswrapper[4840]: E0930 14:47:31.384801 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e1e0905-68d7-4038-91fa-3c9934c5cccc" containerName="collect-profiles" Sep 30 14:47:31 crc kubenswrapper[4840]: I0930 14:47:31.384817 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e1e0905-68d7-4038-91fa-3c9934c5cccc" containerName="collect-profiles" Sep 30 14:47:31 crc kubenswrapper[4840]: I0930 14:47:31.385071 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e1e0905-68d7-4038-91fa-3c9934c5cccc" containerName="collect-profiles" Sep 30 14:47:31 crc kubenswrapper[4840]: I0930 14:47:31.386853 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f5hcj" Sep 30 14:47:31 crc kubenswrapper[4840]: I0930 14:47:31.394577 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f5hcj"] Sep 30 14:47:31 crc kubenswrapper[4840]: I0930 14:47:31.479317 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd5cb64a-3687-4251-9af8-642d443c36a5-catalog-content\") pod \"redhat-marketplace-f5hcj\" (UID: \"cd5cb64a-3687-4251-9af8-642d443c36a5\") " pod="openshift-marketplace/redhat-marketplace-f5hcj" Sep 30 14:47:31 crc kubenswrapper[4840]: I0930 14:47:31.479407 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd5cb64a-3687-4251-9af8-642d443c36a5-utilities\") pod \"redhat-marketplace-f5hcj\" (UID: \"cd5cb64a-3687-4251-9af8-642d443c36a5\") " pod="openshift-marketplace/redhat-marketplace-f5hcj" Sep 30 14:47:31 crc kubenswrapper[4840]: I0930 14:47:31.479515 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rccx\" (UniqueName: \"kubernetes.io/projected/cd5cb64a-3687-4251-9af8-642d443c36a5-kube-api-access-6rccx\") pod \"redhat-marketplace-f5hcj\" (UID: \"cd5cb64a-3687-4251-9af8-642d443c36a5\") " pod="openshift-marketplace/redhat-marketplace-f5hcj" Sep 30 14:47:31 crc kubenswrapper[4840]: I0930 14:47:31.581704 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd5cb64a-3687-4251-9af8-642d443c36a5-utilities\") pod \"redhat-marketplace-f5hcj\" (UID: \"cd5cb64a-3687-4251-9af8-642d443c36a5\") " pod="openshift-marketplace/redhat-marketplace-f5hcj" Sep 30 14:47:31 crc kubenswrapper[4840]: I0930 14:47:31.581838 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rccx\" (UniqueName: \"kubernetes.io/projected/cd5cb64a-3687-4251-9af8-642d443c36a5-kube-api-access-6rccx\") pod \"redhat-marketplace-f5hcj\" (UID: \"cd5cb64a-3687-4251-9af8-642d443c36a5\") " pod="openshift-marketplace/redhat-marketplace-f5hcj" Sep 30 14:47:31 crc kubenswrapper[4840]: I0930 14:47:31.581966 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd5cb64a-3687-4251-9af8-642d443c36a5-catalog-content\") pod \"redhat-marketplace-f5hcj\" (UID: \"cd5cb64a-3687-4251-9af8-642d443c36a5\") " pod="openshift-marketplace/redhat-marketplace-f5hcj" Sep 30 14:47:31 crc kubenswrapper[4840]: I0930 14:47:31.582177 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd5cb64a-3687-4251-9af8-642d443c36a5-utilities\") pod \"redhat-marketplace-f5hcj\" (UID: \"cd5cb64a-3687-4251-9af8-642d443c36a5\") " pod="openshift-marketplace/redhat-marketplace-f5hcj" Sep 30 14:47:31 crc kubenswrapper[4840]: I0930 14:47:31.582398 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd5cb64a-3687-4251-9af8-642d443c36a5-catalog-content\") pod \"redhat-marketplace-f5hcj\" (UID: \"cd5cb64a-3687-4251-9af8-642d443c36a5\") " pod="openshift-marketplace/redhat-marketplace-f5hcj" Sep 30 14:47:31 crc kubenswrapper[4840]: I0930 14:47:31.604589 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rccx\" (UniqueName: \"kubernetes.io/projected/cd5cb64a-3687-4251-9af8-642d443c36a5-kube-api-access-6rccx\") pod \"redhat-marketplace-f5hcj\" (UID: \"cd5cb64a-3687-4251-9af8-642d443c36a5\") " pod="openshift-marketplace/redhat-marketplace-f5hcj" Sep 30 14:47:31 crc kubenswrapper[4840]: I0930 14:47:31.705829 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f5hcj" Sep 30 14:47:32 crc kubenswrapper[4840]: I0930 14:47:32.127622 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f5hcj"] Sep 30 14:47:32 crc kubenswrapper[4840]: I0930 14:47:32.979327 4840 generic.go:334] "Generic (PLEG): container finished" podID="cd5cb64a-3687-4251-9af8-642d443c36a5" containerID="729d527ba407eac963e98398035332057820e4ae73c3d5eef37e652bb4c2b132" exitCode=0 Sep 30 14:47:32 crc kubenswrapper[4840]: I0930 14:47:32.979394 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5hcj" event={"ID":"cd5cb64a-3687-4251-9af8-642d443c36a5","Type":"ContainerDied","Data":"729d527ba407eac963e98398035332057820e4ae73c3d5eef37e652bb4c2b132"} Sep 30 14:47:32 crc kubenswrapper[4840]: I0930 14:47:32.979461 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5hcj" event={"ID":"cd5cb64a-3687-4251-9af8-642d443c36a5","Type":"ContainerStarted","Data":"1efd3a9998964d31703714eefd28501a93faefa36e1fd90988c36ef667104500"} Sep 30 14:47:32 crc kubenswrapper[4840]: I0930 14:47:32.980915 4840 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 14:47:34 crc kubenswrapper[4840]: I0930 14:47:34.018903 4840 generic.go:334] "Generic (PLEG): container finished" podID="cd5cb64a-3687-4251-9af8-642d443c36a5" containerID="c01f8889283036cb0f2e8ac0217661655bfbc7d01633b1ec3c88d31c6f9f9050" exitCode=0 Sep 30 14:47:34 crc kubenswrapper[4840]: I0930 14:47:34.018995 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5hcj" event={"ID":"cd5cb64a-3687-4251-9af8-642d443c36a5","Type":"ContainerDied","Data":"c01f8889283036cb0f2e8ac0217661655bfbc7d01633b1ec3c88d31c6f9f9050"} Sep 30 14:47:35 crc kubenswrapper[4840]: I0930 14:47:35.030092 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5hcj" event={"ID":"cd5cb64a-3687-4251-9af8-642d443c36a5","Type":"ContainerStarted","Data":"bce7174e95ed22706e386c7df64510b140165ebe65e44ed0ca3e9e2674ef07c3"} Sep 30 14:47:35 crc kubenswrapper[4840]: I0930 14:47:35.048433 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-f5hcj" podStartSLOduration=2.370075781 podStartE2EDuration="4.048413456s" podCreationTimestamp="2025-09-30 14:47:31 +0000 UTC" firstStartedPulling="2025-09-30 14:47:32.980692299 +0000 UTC m=+3081.609778722" lastFinishedPulling="2025-09-30 14:47:34.659029974 +0000 UTC m=+3083.288116397" observedRunningTime="2025-09-30 14:47:35.045917463 +0000 UTC m=+3083.675003906" watchObservedRunningTime="2025-09-30 14:47:35.048413456 +0000 UTC m=+3083.677499879" Sep 30 14:47:41 crc kubenswrapper[4840]: I0930 14:47:41.706914 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-f5hcj" Sep 30 14:47:41 crc kubenswrapper[4840]: I0930 14:47:41.707415 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-f5hcj" Sep 30 14:47:41 crc kubenswrapper[4840]: I0930 14:47:41.754114 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-f5hcj" Sep 30 14:47:42 crc kubenswrapper[4840]: I0930 14:47:42.138934 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-f5hcj" Sep 30 14:47:42 crc kubenswrapper[4840]: I0930 14:47:42.180224 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f5hcj"] Sep 30 14:47:44 crc kubenswrapper[4840]: I0930 14:47:44.128740 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-f5hcj" podUID="cd5cb64a-3687-4251-9af8-642d443c36a5" containerName="registry-server" containerID="cri-o://bce7174e95ed22706e386c7df64510b140165ebe65e44ed0ca3e9e2674ef07c3" gracePeriod=2 Sep 30 14:47:44 crc kubenswrapper[4840]: I0930 14:47:44.626426 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f5hcj" Sep 30 14:47:44 crc kubenswrapper[4840]: I0930 14:47:44.724401 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rccx\" (UniqueName: \"kubernetes.io/projected/cd5cb64a-3687-4251-9af8-642d443c36a5-kube-api-access-6rccx\") pod \"cd5cb64a-3687-4251-9af8-642d443c36a5\" (UID: \"cd5cb64a-3687-4251-9af8-642d443c36a5\") " Sep 30 14:47:44 crc kubenswrapper[4840]: I0930 14:47:44.724674 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd5cb64a-3687-4251-9af8-642d443c36a5-utilities\") pod \"cd5cb64a-3687-4251-9af8-642d443c36a5\" (UID: \"cd5cb64a-3687-4251-9af8-642d443c36a5\") " Sep 30 14:47:44 crc kubenswrapper[4840]: I0930 14:47:44.724806 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd5cb64a-3687-4251-9af8-642d443c36a5-catalog-content\") pod \"cd5cb64a-3687-4251-9af8-642d443c36a5\" (UID: \"cd5cb64a-3687-4251-9af8-642d443c36a5\") " Sep 30 14:47:44 crc kubenswrapper[4840]: I0930 14:47:44.725385 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd5cb64a-3687-4251-9af8-642d443c36a5-utilities" (OuterVolumeSpecName: "utilities") pod "cd5cb64a-3687-4251-9af8-642d443c36a5" (UID: "cd5cb64a-3687-4251-9af8-642d443c36a5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:47:44 crc kubenswrapper[4840]: I0930 14:47:44.730034 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd5cb64a-3687-4251-9af8-642d443c36a5-kube-api-access-6rccx" (OuterVolumeSpecName: "kube-api-access-6rccx") pod "cd5cb64a-3687-4251-9af8-642d443c36a5" (UID: "cd5cb64a-3687-4251-9af8-642d443c36a5"). InnerVolumeSpecName "kube-api-access-6rccx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:47:44 crc kubenswrapper[4840]: I0930 14:47:44.737482 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd5cb64a-3687-4251-9af8-642d443c36a5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cd5cb64a-3687-4251-9af8-642d443c36a5" (UID: "cd5cb64a-3687-4251-9af8-642d443c36a5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:47:44 crc kubenswrapper[4840]: I0930 14:47:44.826857 4840 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd5cb64a-3687-4251-9af8-642d443c36a5-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:47:44 crc kubenswrapper[4840]: I0930 14:47:44.826893 4840 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd5cb64a-3687-4251-9af8-642d443c36a5-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:47:44 crc kubenswrapper[4840]: I0930 14:47:44.826904 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rccx\" (UniqueName: \"kubernetes.io/projected/cd5cb64a-3687-4251-9af8-642d443c36a5-kube-api-access-6rccx\") on node \"crc\" DevicePath \"\"" Sep 30 14:47:45 crc kubenswrapper[4840]: I0930 14:47:45.131825 4840 generic.go:334] "Generic (PLEG): container finished" podID="cd5cb64a-3687-4251-9af8-642d443c36a5" containerID="bce7174e95ed22706e386c7df64510b140165ebe65e44ed0ca3e9e2674ef07c3" exitCode=0 Sep 30 14:47:45 crc kubenswrapper[4840]: I0930 14:47:45.131879 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5hcj" event={"ID":"cd5cb64a-3687-4251-9af8-642d443c36a5","Type":"ContainerDied","Data":"bce7174e95ed22706e386c7df64510b140165ebe65e44ed0ca3e9e2674ef07c3"} Sep 30 14:47:45 crc kubenswrapper[4840]: I0930 14:47:45.131916 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5hcj" event={"ID":"cd5cb64a-3687-4251-9af8-642d443c36a5","Type":"ContainerDied","Data":"1efd3a9998964d31703714eefd28501a93faefa36e1fd90988c36ef667104500"} Sep 30 14:47:45 crc kubenswrapper[4840]: I0930 14:47:45.131956 4840 scope.go:117] "RemoveContainer" containerID="bce7174e95ed22706e386c7df64510b140165ebe65e44ed0ca3e9e2674ef07c3" Sep 30 14:47:45 crc kubenswrapper[4840]: I0930 14:47:45.132070 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f5hcj" Sep 30 14:47:45 crc kubenswrapper[4840]: I0930 14:47:45.162893 4840 scope.go:117] "RemoveContainer" containerID="c01f8889283036cb0f2e8ac0217661655bfbc7d01633b1ec3c88d31c6f9f9050" Sep 30 14:47:45 crc kubenswrapper[4840]: I0930 14:47:45.173833 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f5hcj"] Sep 30 14:47:45 crc kubenswrapper[4840]: I0930 14:47:45.183125 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-f5hcj"] Sep 30 14:47:45 crc kubenswrapper[4840]: I0930 14:47:45.185851 4840 scope.go:117] "RemoveContainer" containerID="729d527ba407eac963e98398035332057820e4ae73c3d5eef37e652bb4c2b132" Sep 30 14:47:45 crc kubenswrapper[4840]: I0930 14:47:45.231168 4840 scope.go:117] "RemoveContainer" containerID="bce7174e95ed22706e386c7df64510b140165ebe65e44ed0ca3e9e2674ef07c3" Sep 30 14:47:45 crc kubenswrapper[4840]: E0930 14:47:45.232080 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bce7174e95ed22706e386c7df64510b140165ebe65e44ed0ca3e9e2674ef07c3\": container with ID starting with bce7174e95ed22706e386c7df64510b140165ebe65e44ed0ca3e9e2674ef07c3 not found: ID does not exist" containerID="bce7174e95ed22706e386c7df64510b140165ebe65e44ed0ca3e9e2674ef07c3" Sep 30 14:47:45 crc kubenswrapper[4840]: I0930 14:47:45.232128 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bce7174e95ed22706e386c7df64510b140165ebe65e44ed0ca3e9e2674ef07c3"} err="failed to get container status \"bce7174e95ed22706e386c7df64510b140165ebe65e44ed0ca3e9e2674ef07c3\": rpc error: code = NotFound desc = could not find container \"bce7174e95ed22706e386c7df64510b140165ebe65e44ed0ca3e9e2674ef07c3\": container with ID starting with bce7174e95ed22706e386c7df64510b140165ebe65e44ed0ca3e9e2674ef07c3 not found: ID does not exist" Sep 30 14:47:45 crc kubenswrapper[4840]: I0930 14:47:45.232161 4840 scope.go:117] "RemoveContainer" containerID="c01f8889283036cb0f2e8ac0217661655bfbc7d01633b1ec3c88d31c6f9f9050" Sep 30 14:47:45 crc kubenswrapper[4840]: E0930 14:47:45.235903 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c01f8889283036cb0f2e8ac0217661655bfbc7d01633b1ec3c88d31c6f9f9050\": container with ID starting with c01f8889283036cb0f2e8ac0217661655bfbc7d01633b1ec3c88d31c6f9f9050 not found: ID does not exist" containerID="c01f8889283036cb0f2e8ac0217661655bfbc7d01633b1ec3c88d31c6f9f9050" Sep 30 14:47:45 crc kubenswrapper[4840]: I0930 14:47:45.235942 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c01f8889283036cb0f2e8ac0217661655bfbc7d01633b1ec3c88d31c6f9f9050"} err="failed to get container status \"c01f8889283036cb0f2e8ac0217661655bfbc7d01633b1ec3c88d31c6f9f9050\": rpc error: code = NotFound desc = could not find container \"c01f8889283036cb0f2e8ac0217661655bfbc7d01633b1ec3c88d31c6f9f9050\": container with ID starting with c01f8889283036cb0f2e8ac0217661655bfbc7d01633b1ec3c88d31c6f9f9050 not found: ID does not exist" Sep 30 14:47:45 crc kubenswrapper[4840]: I0930 14:47:45.235964 4840 scope.go:117] "RemoveContainer" containerID="729d527ba407eac963e98398035332057820e4ae73c3d5eef37e652bb4c2b132" Sep 30 14:47:45 crc kubenswrapper[4840]: E0930 14:47:45.236332 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"729d527ba407eac963e98398035332057820e4ae73c3d5eef37e652bb4c2b132\": container with ID starting with 729d527ba407eac963e98398035332057820e4ae73c3d5eef37e652bb4c2b132 not found: ID does not exist" containerID="729d527ba407eac963e98398035332057820e4ae73c3d5eef37e652bb4c2b132" Sep 30 14:47:45 crc kubenswrapper[4840]: I0930 14:47:45.236355 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"729d527ba407eac963e98398035332057820e4ae73c3d5eef37e652bb4c2b132"} err="failed to get container status \"729d527ba407eac963e98398035332057820e4ae73c3d5eef37e652bb4c2b132\": rpc error: code = NotFound desc = could not find container \"729d527ba407eac963e98398035332057820e4ae73c3d5eef37e652bb4c2b132\": container with ID starting with 729d527ba407eac963e98398035332057820e4ae73c3d5eef37e652bb4c2b132 not found: ID does not exist" Sep 30 14:47:46 crc kubenswrapper[4840]: I0930 14:47:46.126338 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd5cb64a-3687-4251-9af8-642d443c36a5" path="/var/lib/kubelet/pods/cd5cb64a-3687-4251-9af8-642d443c36a5/volumes" Sep 30 14:47:51 crc kubenswrapper[4840]: I0930 14:47:51.871635 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:47:51 crc kubenswrapper[4840]: I0930 14:47:51.872253 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:48:21 crc kubenswrapper[4840]: I0930 14:48:21.871687 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:48:21 crc kubenswrapper[4840]: I0930 14:48:21.872301 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:48:21 crc kubenswrapper[4840]: I0930 14:48:21.872356 4840 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-747gk" Sep 30 14:48:21 crc kubenswrapper[4840]: I0930 14:48:21.873120 4840 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bb42e5820a8aaddbc79e0404192d904aa52fd67a0816b292b3605d7cf3a9d5b0"} pod="openshift-machine-config-operator/machine-config-daemon-747gk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 14:48:21 crc kubenswrapper[4840]: I0930 14:48:21.873196 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" containerID="cri-o://bb42e5820a8aaddbc79e0404192d904aa52fd67a0816b292b3605d7cf3a9d5b0" gracePeriod=600 Sep 30 14:48:22 crc kubenswrapper[4840]: E0930 14:48:22.017664 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:48:22 crc kubenswrapper[4840]: I0930 14:48:22.462841 4840 generic.go:334] "Generic (PLEG): container finished" podID="10e8b890-7f20-4a36-8e03-898620cf599a" containerID="bb42e5820a8aaddbc79e0404192d904aa52fd67a0816b292b3605d7cf3a9d5b0" exitCode=0 Sep 30 14:48:22 crc kubenswrapper[4840]: I0930 14:48:22.463041 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" event={"ID":"10e8b890-7f20-4a36-8e03-898620cf599a","Type":"ContainerDied","Data":"bb42e5820a8aaddbc79e0404192d904aa52fd67a0816b292b3605d7cf3a9d5b0"} Sep 30 14:48:22 crc kubenswrapper[4840]: I0930 14:48:22.463333 4840 scope.go:117] "RemoveContainer" containerID="114f6accf72be8011201b200dc40143e1ba0f13c3f60c2a725c667df5ceff985" Sep 30 14:48:22 crc kubenswrapper[4840]: I0930 14:48:22.464096 4840 scope.go:117] "RemoveContainer" containerID="bb42e5820a8aaddbc79e0404192d904aa52fd67a0816b292b3605d7cf3a9d5b0" Sep 30 14:48:22 crc kubenswrapper[4840]: E0930 14:48:22.464641 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:48:34 crc kubenswrapper[4840]: I0930 14:48:34.117513 4840 scope.go:117] "RemoveContainer" containerID="bb42e5820a8aaddbc79e0404192d904aa52fd67a0816b292b3605d7cf3a9d5b0" Sep 30 14:48:34 crc kubenswrapper[4840]: E0930 14:48:34.118438 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:48:47 crc kubenswrapper[4840]: I0930 14:48:47.116284 4840 scope.go:117] "RemoveContainer" containerID="bb42e5820a8aaddbc79e0404192d904aa52fd67a0816b292b3605d7cf3a9d5b0" Sep 30 14:48:47 crc kubenswrapper[4840]: E0930 14:48:47.117137 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:49:00 crc kubenswrapper[4840]: I0930 14:49:00.116534 4840 scope.go:117] "RemoveContainer" containerID="bb42e5820a8aaddbc79e0404192d904aa52fd67a0816b292b3605d7cf3a9d5b0" Sep 30 14:49:00 crc kubenswrapper[4840]: E0930 14:49:00.117360 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:49:02 crc kubenswrapper[4840]: I0930 14:49:02.599827 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bbwsx"] Sep 30 14:49:02 crc kubenswrapper[4840]: E0930 14:49:02.600816 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd5cb64a-3687-4251-9af8-642d443c36a5" containerName="extract-content" Sep 30 14:49:02 crc kubenswrapper[4840]: I0930 14:49:02.600830 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd5cb64a-3687-4251-9af8-642d443c36a5" containerName="extract-content" Sep 30 14:49:02 crc kubenswrapper[4840]: E0930 14:49:02.600847 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd5cb64a-3687-4251-9af8-642d443c36a5" containerName="extract-utilities" Sep 30 14:49:02 crc kubenswrapper[4840]: I0930 14:49:02.600853 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd5cb64a-3687-4251-9af8-642d443c36a5" containerName="extract-utilities" Sep 30 14:49:02 crc kubenswrapper[4840]: E0930 14:49:02.600866 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd5cb64a-3687-4251-9af8-642d443c36a5" containerName="registry-server" Sep 30 14:49:02 crc kubenswrapper[4840]: I0930 14:49:02.600872 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd5cb64a-3687-4251-9af8-642d443c36a5" containerName="registry-server" Sep 30 14:49:02 crc kubenswrapper[4840]: I0930 14:49:02.601069 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd5cb64a-3687-4251-9af8-642d443c36a5" containerName="registry-server" Sep 30 14:49:02 crc kubenswrapper[4840]: I0930 14:49:02.602528 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bbwsx" Sep 30 14:49:02 crc kubenswrapper[4840]: I0930 14:49:02.612627 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bbwsx"] Sep 30 14:49:02 crc kubenswrapper[4840]: I0930 14:49:02.699682 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvk5v\" (UniqueName: \"kubernetes.io/projected/b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0-kube-api-access-dvk5v\") pod \"redhat-operators-bbwsx\" (UID: \"b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0\") " pod="openshift-marketplace/redhat-operators-bbwsx" Sep 30 14:49:02 crc kubenswrapper[4840]: I0930 14:49:02.699984 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0-utilities\") pod \"redhat-operators-bbwsx\" (UID: \"b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0\") " pod="openshift-marketplace/redhat-operators-bbwsx" Sep 30 14:49:02 crc kubenswrapper[4840]: I0930 14:49:02.700107 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0-catalog-content\") pod \"redhat-operators-bbwsx\" (UID: \"b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0\") " pod="openshift-marketplace/redhat-operators-bbwsx" Sep 30 14:49:02 crc kubenswrapper[4840]: I0930 14:49:02.801635 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0-utilities\") pod \"redhat-operators-bbwsx\" (UID: \"b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0\") " pod="openshift-marketplace/redhat-operators-bbwsx" Sep 30 14:49:02 crc kubenswrapper[4840]: I0930 14:49:02.801735 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0-catalog-content\") pod \"redhat-operators-bbwsx\" (UID: \"b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0\") " pod="openshift-marketplace/redhat-operators-bbwsx" Sep 30 14:49:02 crc kubenswrapper[4840]: I0930 14:49:02.801833 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvk5v\" (UniqueName: \"kubernetes.io/projected/b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0-kube-api-access-dvk5v\") pod \"redhat-operators-bbwsx\" (UID: \"b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0\") " pod="openshift-marketplace/redhat-operators-bbwsx" Sep 30 14:49:02 crc kubenswrapper[4840]: I0930 14:49:02.802415 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0-utilities\") pod \"redhat-operators-bbwsx\" (UID: \"b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0\") " pod="openshift-marketplace/redhat-operators-bbwsx" Sep 30 14:49:02 crc kubenswrapper[4840]: I0930 14:49:02.802436 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0-catalog-content\") pod \"redhat-operators-bbwsx\" (UID: \"b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0\") " pod="openshift-marketplace/redhat-operators-bbwsx" Sep 30 14:49:02 crc kubenswrapper[4840]: I0930 14:49:02.836835 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvk5v\" (UniqueName: \"kubernetes.io/projected/b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0-kube-api-access-dvk5v\") pod \"redhat-operators-bbwsx\" (UID: \"b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0\") " pod="openshift-marketplace/redhat-operators-bbwsx" Sep 30 14:49:02 crc kubenswrapper[4840]: I0930 14:49:02.942802 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bbwsx" Sep 30 14:49:03 crc kubenswrapper[4840]: I0930 14:49:03.419247 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bbwsx"] Sep 30 14:49:03 crc kubenswrapper[4840]: I0930 14:49:03.845051 4840 generic.go:334] "Generic (PLEG): container finished" podID="b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0" containerID="4fdc3926858b119cc33017e0068cdc3a1fffab7a19b27331eeae7ebdd2bf00ab" exitCode=0 Sep 30 14:49:03 crc kubenswrapper[4840]: I0930 14:49:03.845621 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bbwsx" event={"ID":"b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0","Type":"ContainerDied","Data":"4fdc3926858b119cc33017e0068cdc3a1fffab7a19b27331eeae7ebdd2bf00ab"} Sep 30 14:49:03 crc kubenswrapper[4840]: I0930 14:49:03.845697 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bbwsx" event={"ID":"b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0","Type":"ContainerStarted","Data":"3c1a4f96ac3baa89339810d6f5ed3433d517670e3eb6c8b8659838a49def3991"} Sep 30 14:49:06 crc kubenswrapper[4840]: I0930 14:49:06.876167 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bbwsx" event={"ID":"b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0","Type":"ContainerStarted","Data":"16be081b29263c0ca48bf197c4584c627b9c660b98fc71027d6bab8361653df5"} Sep 30 14:49:10 crc kubenswrapper[4840]: I0930 14:49:10.919495 4840 generic.go:334] "Generic (PLEG): container finished" podID="b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0" containerID="16be081b29263c0ca48bf197c4584c627b9c660b98fc71027d6bab8361653df5" exitCode=0 Sep 30 14:49:10 crc kubenswrapper[4840]: I0930 14:49:10.919647 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bbwsx" event={"ID":"b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0","Type":"ContainerDied","Data":"16be081b29263c0ca48bf197c4584c627b9c660b98fc71027d6bab8361653df5"} Sep 30 14:49:13 crc kubenswrapper[4840]: I0930 14:49:13.117407 4840 scope.go:117] "RemoveContainer" containerID="bb42e5820a8aaddbc79e0404192d904aa52fd67a0816b292b3605d7cf3a9d5b0" Sep 30 14:49:13 crc kubenswrapper[4840]: E0930 14:49:13.118688 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:49:13 crc kubenswrapper[4840]: I0930 14:49:13.947239 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bbwsx" event={"ID":"b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0","Type":"ContainerStarted","Data":"1b5dc8d0b4ee067bf698fa27177a1409ba9abd32529d09a7d09add807aa65416"} Sep 30 14:49:15 crc kubenswrapper[4840]: I0930 14:49:15.009068 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bbwsx" podStartSLOduration=3.33439996 podStartE2EDuration="13.009048299s" podCreationTimestamp="2025-09-30 14:49:02 +0000 UTC" firstStartedPulling="2025-09-30 14:49:03.85069983 +0000 UTC m=+3172.479786253" lastFinishedPulling="2025-09-30 14:49:13.525348169 +0000 UTC m=+3182.154434592" observedRunningTime="2025-09-30 14:49:15.000186423 +0000 UTC m=+3183.629272896" watchObservedRunningTime="2025-09-30 14:49:15.009048299 +0000 UTC m=+3183.638134722" Sep 30 14:49:22 crc kubenswrapper[4840]: I0930 14:49:22.943163 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bbwsx" Sep 30 14:49:22 crc kubenswrapper[4840]: I0930 14:49:22.943863 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bbwsx" Sep 30 14:49:23 crc kubenswrapper[4840]: I0930 14:49:23.003358 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bbwsx" Sep 30 14:49:23 crc kubenswrapper[4840]: I0930 14:49:23.071963 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bbwsx" Sep 30 14:49:23 crc kubenswrapper[4840]: I0930 14:49:23.237343 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bbwsx"] Sep 30 14:49:25 crc kubenswrapper[4840]: I0930 14:49:25.047829 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bbwsx" podUID="b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0" containerName="registry-server" containerID="cri-o://1b5dc8d0b4ee067bf698fa27177a1409ba9abd32529d09a7d09add807aa65416" gracePeriod=2 Sep 30 14:49:25 crc kubenswrapper[4840]: I0930 14:49:25.117151 4840 scope.go:117] "RemoveContainer" containerID="bb42e5820a8aaddbc79e0404192d904aa52fd67a0816b292b3605d7cf3a9d5b0" Sep 30 14:49:25 crc kubenswrapper[4840]: E0930 14:49:25.117904 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:49:25 crc kubenswrapper[4840]: I0930 14:49:25.534152 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bbwsx" Sep 30 14:49:25 crc kubenswrapper[4840]: I0930 14:49:25.645327 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0-utilities\") pod \"b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0\" (UID: \"b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0\") " Sep 30 14:49:25 crc kubenswrapper[4840]: I0930 14:49:25.645376 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0-catalog-content\") pod \"b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0\" (UID: \"b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0\") " Sep 30 14:49:25 crc kubenswrapper[4840]: I0930 14:49:25.645529 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvk5v\" (UniqueName: \"kubernetes.io/projected/b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0-kube-api-access-dvk5v\") pod \"b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0\" (UID: \"b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0\") " Sep 30 14:49:25 crc kubenswrapper[4840]: I0930 14:49:25.646175 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0-utilities" (OuterVolumeSpecName: "utilities") pod "b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0" (UID: "b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:49:25 crc kubenswrapper[4840]: I0930 14:49:25.650708 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0-kube-api-access-dvk5v" (OuterVolumeSpecName: "kube-api-access-dvk5v") pod "b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0" (UID: "b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0"). InnerVolumeSpecName "kube-api-access-dvk5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:49:25 crc kubenswrapper[4840]: I0930 14:49:25.747842 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvk5v\" (UniqueName: \"kubernetes.io/projected/b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0-kube-api-access-dvk5v\") on node \"crc\" DevicePath \"\"" Sep 30 14:49:25 crc kubenswrapper[4840]: I0930 14:49:25.747877 4840 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:49:25 crc kubenswrapper[4840]: I0930 14:49:25.756172 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0" (UID: "b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:49:25 crc kubenswrapper[4840]: I0930 14:49:25.849511 4840 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:49:26 crc kubenswrapper[4840]: I0930 14:49:26.057632 4840 generic.go:334] "Generic (PLEG): container finished" podID="b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0" containerID="1b5dc8d0b4ee067bf698fa27177a1409ba9abd32529d09a7d09add807aa65416" exitCode=0 Sep 30 14:49:26 crc kubenswrapper[4840]: I0930 14:49:26.057690 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bbwsx" event={"ID":"b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0","Type":"ContainerDied","Data":"1b5dc8d0b4ee067bf698fa27177a1409ba9abd32529d09a7d09add807aa65416"} Sep 30 14:49:26 crc kubenswrapper[4840]: I0930 14:49:26.057721 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bbwsx" event={"ID":"b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0","Type":"ContainerDied","Data":"3c1a4f96ac3baa89339810d6f5ed3433d517670e3eb6c8b8659838a49def3991"} Sep 30 14:49:26 crc kubenswrapper[4840]: I0930 14:49:26.057740 4840 scope.go:117] "RemoveContainer" containerID="1b5dc8d0b4ee067bf698fa27177a1409ba9abd32529d09a7d09add807aa65416" Sep 30 14:49:26 crc kubenswrapper[4840]: I0930 14:49:26.057776 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bbwsx" Sep 30 14:49:26 crc kubenswrapper[4840]: I0930 14:49:26.081437 4840 scope.go:117] "RemoveContainer" containerID="16be081b29263c0ca48bf197c4584c627b9c660b98fc71027d6bab8361653df5" Sep 30 14:49:26 crc kubenswrapper[4840]: I0930 14:49:26.096935 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bbwsx"] Sep 30 14:49:26 crc kubenswrapper[4840]: I0930 14:49:26.106738 4840 scope.go:117] "RemoveContainer" containerID="4fdc3926858b119cc33017e0068cdc3a1fffab7a19b27331eeae7ebdd2bf00ab" Sep 30 14:49:26 crc kubenswrapper[4840]: I0930 14:49:26.107818 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bbwsx"] Sep 30 14:49:26 crc kubenswrapper[4840]: I0930 14:49:26.133237 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0" path="/var/lib/kubelet/pods/b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0/volumes" Sep 30 14:49:26 crc kubenswrapper[4840]: I0930 14:49:26.147907 4840 scope.go:117] "RemoveContainer" containerID="1b5dc8d0b4ee067bf698fa27177a1409ba9abd32529d09a7d09add807aa65416" Sep 30 14:49:26 crc kubenswrapper[4840]: E0930 14:49:26.149643 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b5dc8d0b4ee067bf698fa27177a1409ba9abd32529d09a7d09add807aa65416\": container with ID starting with 1b5dc8d0b4ee067bf698fa27177a1409ba9abd32529d09a7d09add807aa65416 not found: ID does not exist" containerID="1b5dc8d0b4ee067bf698fa27177a1409ba9abd32529d09a7d09add807aa65416" Sep 30 14:49:26 crc kubenswrapper[4840]: I0930 14:49:26.149678 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b5dc8d0b4ee067bf698fa27177a1409ba9abd32529d09a7d09add807aa65416"} err="failed to get container status \"1b5dc8d0b4ee067bf698fa27177a1409ba9abd32529d09a7d09add807aa65416\": rpc error: code = NotFound desc = could not find container \"1b5dc8d0b4ee067bf698fa27177a1409ba9abd32529d09a7d09add807aa65416\": container with ID starting with 1b5dc8d0b4ee067bf698fa27177a1409ba9abd32529d09a7d09add807aa65416 not found: ID does not exist" Sep 30 14:49:26 crc kubenswrapper[4840]: I0930 14:49:26.149698 4840 scope.go:117] "RemoveContainer" containerID="16be081b29263c0ca48bf197c4584c627b9c660b98fc71027d6bab8361653df5" Sep 30 14:49:26 crc kubenswrapper[4840]: E0930 14:49:26.151231 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16be081b29263c0ca48bf197c4584c627b9c660b98fc71027d6bab8361653df5\": container with ID starting with 16be081b29263c0ca48bf197c4584c627b9c660b98fc71027d6bab8361653df5 not found: ID does not exist" containerID="16be081b29263c0ca48bf197c4584c627b9c660b98fc71027d6bab8361653df5" Sep 30 14:49:26 crc kubenswrapper[4840]: I0930 14:49:26.151258 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16be081b29263c0ca48bf197c4584c627b9c660b98fc71027d6bab8361653df5"} err="failed to get container status \"16be081b29263c0ca48bf197c4584c627b9c660b98fc71027d6bab8361653df5\": rpc error: code = NotFound desc = could not find container \"16be081b29263c0ca48bf197c4584c627b9c660b98fc71027d6bab8361653df5\": container with ID starting with 16be081b29263c0ca48bf197c4584c627b9c660b98fc71027d6bab8361653df5 not found: ID does not exist" Sep 30 14:49:26 crc kubenswrapper[4840]: I0930 14:49:26.151275 4840 scope.go:117] "RemoveContainer" containerID="4fdc3926858b119cc33017e0068cdc3a1fffab7a19b27331eeae7ebdd2bf00ab" Sep 30 14:49:26 crc kubenswrapper[4840]: E0930 14:49:26.151541 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4fdc3926858b119cc33017e0068cdc3a1fffab7a19b27331eeae7ebdd2bf00ab\": container with ID starting with 4fdc3926858b119cc33017e0068cdc3a1fffab7a19b27331eeae7ebdd2bf00ab not found: ID does not exist" containerID="4fdc3926858b119cc33017e0068cdc3a1fffab7a19b27331eeae7ebdd2bf00ab" Sep 30 14:49:26 crc kubenswrapper[4840]: I0930 14:49:26.151589 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fdc3926858b119cc33017e0068cdc3a1fffab7a19b27331eeae7ebdd2bf00ab"} err="failed to get container status \"4fdc3926858b119cc33017e0068cdc3a1fffab7a19b27331eeae7ebdd2bf00ab\": rpc error: code = NotFound desc = could not find container \"4fdc3926858b119cc33017e0068cdc3a1fffab7a19b27331eeae7ebdd2bf00ab\": container with ID starting with 4fdc3926858b119cc33017e0068cdc3a1fffab7a19b27331eeae7ebdd2bf00ab not found: ID does not exist" Sep 30 14:49:39 crc kubenswrapper[4840]: I0930 14:49:39.116617 4840 scope.go:117] "RemoveContainer" containerID="bb42e5820a8aaddbc79e0404192d904aa52fd67a0816b292b3605d7cf3a9d5b0" Sep 30 14:49:39 crc kubenswrapper[4840]: E0930 14:49:39.117767 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:49:51 crc kubenswrapper[4840]: I0930 14:49:51.117055 4840 scope.go:117] "RemoveContainer" containerID="bb42e5820a8aaddbc79e0404192d904aa52fd67a0816b292b3605d7cf3a9d5b0" Sep 30 14:49:51 crc kubenswrapper[4840]: E0930 14:49:51.117864 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:50:03 crc kubenswrapper[4840]: I0930 14:50:03.116372 4840 scope.go:117] "RemoveContainer" containerID="bb42e5820a8aaddbc79e0404192d904aa52fd67a0816b292b3605d7cf3a9d5b0" Sep 30 14:50:03 crc kubenswrapper[4840]: E0930 14:50:03.117132 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:50:18 crc kubenswrapper[4840]: I0930 14:50:18.117049 4840 scope.go:117] "RemoveContainer" containerID="bb42e5820a8aaddbc79e0404192d904aa52fd67a0816b292b3605d7cf3a9d5b0" Sep 30 14:50:18 crc kubenswrapper[4840]: E0930 14:50:18.117843 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:50:32 crc kubenswrapper[4840]: I0930 14:50:32.125179 4840 scope.go:117] "RemoveContainer" containerID="bb42e5820a8aaddbc79e0404192d904aa52fd67a0816b292b3605d7cf3a9d5b0" Sep 30 14:50:32 crc kubenswrapper[4840]: E0930 14:50:32.126017 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:50:38 crc kubenswrapper[4840]: I0930 14:50:38.152762 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9z6ms"] Sep 30 14:50:38 crc kubenswrapper[4840]: E0930 14:50:38.154873 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0" containerName="registry-server" Sep 30 14:50:38 crc kubenswrapper[4840]: I0930 14:50:38.154900 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0" containerName="registry-server" Sep 30 14:50:38 crc kubenswrapper[4840]: E0930 14:50:38.154945 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0" containerName="extract-content" Sep 30 14:50:38 crc kubenswrapper[4840]: I0930 14:50:38.154957 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0" containerName="extract-content" Sep 30 14:50:38 crc kubenswrapper[4840]: E0930 14:50:38.154982 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0" containerName="extract-utilities" Sep 30 14:50:38 crc kubenswrapper[4840]: I0930 14:50:38.154996 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0" containerName="extract-utilities" Sep 30 14:50:38 crc kubenswrapper[4840]: I0930 14:50:38.155404 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="b38f0a2f-a2c6-4c40-a468-7e7e1f7598f0" containerName="registry-server" Sep 30 14:50:38 crc kubenswrapper[4840]: I0930 14:50:38.158015 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9z6ms" Sep 30 14:50:38 crc kubenswrapper[4840]: I0930 14:50:38.163931 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9z6ms"] Sep 30 14:50:38 crc kubenswrapper[4840]: I0930 14:50:38.274710 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gb4sz\" (UniqueName: \"kubernetes.io/projected/6c84799c-2adb-492a-a005-509b4da94289-kube-api-access-gb4sz\") pod \"certified-operators-9z6ms\" (UID: \"6c84799c-2adb-492a-a005-509b4da94289\") " pod="openshift-marketplace/certified-operators-9z6ms" Sep 30 14:50:38 crc kubenswrapper[4840]: I0930 14:50:38.274889 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c84799c-2adb-492a-a005-509b4da94289-utilities\") pod \"certified-operators-9z6ms\" (UID: \"6c84799c-2adb-492a-a005-509b4da94289\") " pod="openshift-marketplace/certified-operators-9z6ms" Sep 30 14:50:38 crc kubenswrapper[4840]: I0930 14:50:38.275058 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c84799c-2adb-492a-a005-509b4da94289-catalog-content\") pod \"certified-operators-9z6ms\" (UID: \"6c84799c-2adb-492a-a005-509b4da94289\") " pod="openshift-marketplace/certified-operators-9z6ms" Sep 30 14:50:38 crc kubenswrapper[4840]: I0930 14:50:38.377142 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gb4sz\" (UniqueName: \"kubernetes.io/projected/6c84799c-2adb-492a-a005-509b4da94289-kube-api-access-gb4sz\") pod \"certified-operators-9z6ms\" (UID: \"6c84799c-2adb-492a-a005-509b4da94289\") " pod="openshift-marketplace/certified-operators-9z6ms" Sep 30 14:50:38 crc kubenswrapper[4840]: I0930 14:50:38.377437 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c84799c-2adb-492a-a005-509b4da94289-utilities\") pod \"certified-operators-9z6ms\" (UID: \"6c84799c-2adb-492a-a005-509b4da94289\") " pod="openshift-marketplace/certified-operators-9z6ms" Sep 30 14:50:38 crc kubenswrapper[4840]: I0930 14:50:38.377486 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c84799c-2adb-492a-a005-509b4da94289-catalog-content\") pod \"certified-operators-9z6ms\" (UID: \"6c84799c-2adb-492a-a005-509b4da94289\") " pod="openshift-marketplace/certified-operators-9z6ms" Sep 30 14:50:38 crc kubenswrapper[4840]: I0930 14:50:38.377978 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c84799c-2adb-492a-a005-509b4da94289-catalog-content\") pod \"certified-operators-9z6ms\" (UID: \"6c84799c-2adb-492a-a005-509b4da94289\") " pod="openshift-marketplace/certified-operators-9z6ms" Sep 30 14:50:38 crc kubenswrapper[4840]: I0930 14:50:38.378198 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c84799c-2adb-492a-a005-509b4da94289-utilities\") pod \"certified-operators-9z6ms\" (UID: \"6c84799c-2adb-492a-a005-509b4da94289\") " pod="openshift-marketplace/certified-operators-9z6ms" Sep 30 14:50:38 crc kubenswrapper[4840]: I0930 14:50:38.415779 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gb4sz\" (UniqueName: \"kubernetes.io/projected/6c84799c-2adb-492a-a005-509b4da94289-kube-api-access-gb4sz\") pod \"certified-operators-9z6ms\" (UID: \"6c84799c-2adb-492a-a005-509b4da94289\") " pod="openshift-marketplace/certified-operators-9z6ms" Sep 30 14:50:38 crc kubenswrapper[4840]: I0930 14:50:38.488256 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9z6ms" Sep 30 14:50:38 crc kubenswrapper[4840]: I0930 14:50:38.971762 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9z6ms"] Sep 30 14:50:39 crc kubenswrapper[4840]: I0930 14:50:39.708312 4840 generic.go:334] "Generic (PLEG): container finished" podID="6c84799c-2adb-492a-a005-509b4da94289" containerID="94f28a3e0f87efd2145fe14edfeaeb4a285086eb057b2c45c844a5c4d70a0d64" exitCode=0 Sep 30 14:50:39 crc kubenswrapper[4840]: I0930 14:50:39.708399 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9z6ms" event={"ID":"6c84799c-2adb-492a-a005-509b4da94289","Type":"ContainerDied","Data":"94f28a3e0f87efd2145fe14edfeaeb4a285086eb057b2c45c844a5c4d70a0d64"} Sep 30 14:50:39 crc kubenswrapper[4840]: I0930 14:50:39.708611 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9z6ms" event={"ID":"6c84799c-2adb-492a-a005-509b4da94289","Type":"ContainerStarted","Data":"88fb2b9e6b9c11a9e92f2f5bf9dee6690d305fb93ac71f3a4eaf22643ec38f49"} Sep 30 14:50:40 crc kubenswrapper[4840]: I0930 14:50:40.717918 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9z6ms" event={"ID":"6c84799c-2adb-492a-a005-509b4da94289","Type":"ContainerStarted","Data":"29d010a97caee4897176d4184ebd06c9389a114774136117a85bdadce370bfd4"} Sep 30 14:50:41 crc kubenswrapper[4840]: I0930 14:50:41.727712 4840 generic.go:334] "Generic (PLEG): container finished" podID="6c84799c-2adb-492a-a005-509b4da94289" containerID="29d010a97caee4897176d4184ebd06c9389a114774136117a85bdadce370bfd4" exitCode=0 Sep 30 14:50:41 crc kubenswrapper[4840]: I0930 14:50:41.727763 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9z6ms" event={"ID":"6c84799c-2adb-492a-a005-509b4da94289","Type":"ContainerDied","Data":"29d010a97caee4897176d4184ebd06c9389a114774136117a85bdadce370bfd4"} Sep 30 14:50:42 crc kubenswrapper[4840]: I0930 14:50:42.739123 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9z6ms" event={"ID":"6c84799c-2adb-492a-a005-509b4da94289","Type":"ContainerStarted","Data":"959f50719ef40add2c08eaa399cd8fe1c83cb7fd69ec244b942007c83407ffc8"} Sep 30 14:50:42 crc kubenswrapper[4840]: I0930 14:50:42.762359 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9z6ms" podStartSLOduration=2.338475489 podStartE2EDuration="4.762335189s" podCreationTimestamp="2025-09-30 14:50:38 +0000 UTC" firstStartedPulling="2025-09-30 14:50:39.711468574 +0000 UTC m=+3268.340555037" lastFinishedPulling="2025-09-30 14:50:42.135328314 +0000 UTC m=+3270.764414737" observedRunningTime="2025-09-30 14:50:42.755490974 +0000 UTC m=+3271.384577417" watchObservedRunningTime="2025-09-30 14:50:42.762335189 +0000 UTC m=+3271.391421622" Sep 30 14:50:43 crc kubenswrapper[4840]: I0930 14:50:43.117255 4840 scope.go:117] "RemoveContainer" containerID="bb42e5820a8aaddbc79e0404192d904aa52fd67a0816b292b3605d7cf3a9d5b0" Sep 30 14:50:43 crc kubenswrapper[4840]: E0930 14:50:43.117679 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:50:48 crc kubenswrapper[4840]: I0930 14:50:48.489190 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9z6ms" Sep 30 14:50:48 crc kubenswrapper[4840]: I0930 14:50:48.491005 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9z6ms" Sep 30 14:50:48 crc kubenswrapper[4840]: I0930 14:50:48.536240 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9z6ms" Sep 30 14:50:48 crc kubenswrapper[4840]: I0930 14:50:48.839648 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9z6ms" Sep 30 14:50:48 crc kubenswrapper[4840]: I0930 14:50:48.901844 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9z6ms"] Sep 30 14:50:50 crc kubenswrapper[4840]: I0930 14:50:50.813687 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9z6ms" podUID="6c84799c-2adb-492a-a005-509b4da94289" containerName="registry-server" containerID="cri-o://959f50719ef40add2c08eaa399cd8fe1c83cb7fd69ec244b942007c83407ffc8" gracePeriod=2 Sep 30 14:50:51 crc kubenswrapper[4840]: I0930 14:50:51.366900 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9z6ms" Sep 30 14:50:51 crc kubenswrapper[4840]: I0930 14:50:51.429669 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gb4sz\" (UniqueName: \"kubernetes.io/projected/6c84799c-2adb-492a-a005-509b4da94289-kube-api-access-gb4sz\") pod \"6c84799c-2adb-492a-a005-509b4da94289\" (UID: \"6c84799c-2adb-492a-a005-509b4da94289\") " Sep 30 14:50:51 crc kubenswrapper[4840]: I0930 14:50:51.430006 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c84799c-2adb-492a-a005-509b4da94289-catalog-content\") pod \"6c84799c-2adb-492a-a005-509b4da94289\" (UID: \"6c84799c-2adb-492a-a005-509b4da94289\") " Sep 30 14:50:51 crc kubenswrapper[4840]: I0930 14:50:51.430171 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c84799c-2adb-492a-a005-509b4da94289-utilities\") pod \"6c84799c-2adb-492a-a005-509b4da94289\" (UID: \"6c84799c-2adb-492a-a005-509b4da94289\") " Sep 30 14:50:51 crc kubenswrapper[4840]: I0930 14:50:51.430834 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c84799c-2adb-492a-a005-509b4da94289-utilities" (OuterVolumeSpecName: "utilities") pod "6c84799c-2adb-492a-a005-509b4da94289" (UID: "6c84799c-2adb-492a-a005-509b4da94289"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:50:51 crc kubenswrapper[4840]: I0930 14:50:51.437685 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c84799c-2adb-492a-a005-509b4da94289-kube-api-access-gb4sz" (OuterVolumeSpecName: "kube-api-access-gb4sz") pod "6c84799c-2adb-492a-a005-509b4da94289" (UID: "6c84799c-2adb-492a-a005-509b4da94289"). InnerVolumeSpecName "kube-api-access-gb4sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:50:51 crc kubenswrapper[4840]: I0930 14:50:51.521145 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c84799c-2adb-492a-a005-509b4da94289-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6c84799c-2adb-492a-a005-509b4da94289" (UID: "6c84799c-2adb-492a-a005-509b4da94289"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:50:51 crc kubenswrapper[4840]: I0930 14:50:51.532805 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gb4sz\" (UniqueName: \"kubernetes.io/projected/6c84799c-2adb-492a-a005-509b4da94289-kube-api-access-gb4sz\") on node \"crc\" DevicePath \"\"" Sep 30 14:50:51 crc kubenswrapper[4840]: I0930 14:50:51.532849 4840 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c84799c-2adb-492a-a005-509b4da94289-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:50:51 crc kubenswrapper[4840]: I0930 14:50:51.532863 4840 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c84799c-2adb-492a-a005-509b4da94289-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:50:51 crc kubenswrapper[4840]: I0930 14:50:51.824342 4840 generic.go:334] "Generic (PLEG): container finished" podID="6c84799c-2adb-492a-a005-509b4da94289" containerID="959f50719ef40add2c08eaa399cd8fe1c83cb7fd69ec244b942007c83407ffc8" exitCode=0 Sep 30 14:50:51 crc kubenswrapper[4840]: I0930 14:50:51.824404 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9z6ms" Sep 30 14:50:51 crc kubenswrapper[4840]: I0930 14:50:51.824410 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9z6ms" event={"ID":"6c84799c-2adb-492a-a005-509b4da94289","Type":"ContainerDied","Data":"959f50719ef40add2c08eaa399cd8fe1c83cb7fd69ec244b942007c83407ffc8"} Sep 30 14:50:51 crc kubenswrapper[4840]: I0930 14:50:51.824797 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9z6ms" event={"ID":"6c84799c-2adb-492a-a005-509b4da94289","Type":"ContainerDied","Data":"88fb2b9e6b9c11a9e92f2f5bf9dee6690d305fb93ac71f3a4eaf22643ec38f49"} Sep 30 14:50:51 crc kubenswrapper[4840]: I0930 14:50:51.824819 4840 scope.go:117] "RemoveContainer" containerID="959f50719ef40add2c08eaa399cd8fe1c83cb7fd69ec244b942007c83407ffc8" Sep 30 14:50:51 crc kubenswrapper[4840]: I0930 14:50:51.845465 4840 scope.go:117] "RemoveContainer" containerID="29d010a97caee4897176d4184ebd06c9389a114774136117a85bdadce370bfd4" Sep 30 14:50:51 crc kubenswrapper[4840]: I0930 14:50:51.866463 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9z6ms"] Sep 30 14:50:51 crc kubenswrapper[4840]: I0930 14:50:51.877299 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9z6ms"] Sep 30 14:50:51 crc kubenswrapper[4840]: I0930 14:50:51.882678 4840 scope.go:117] "RemoveContainer" containerID="94f28a3e0f87efd2145fe14edfeaeb4a285086eb057b2c45c844a5c4d70a0d64" Sep 30 14:50:51 crc kubenswrapper[4840]: I0930 14:50:51.911373 4840 scope.go:117] "RemoveContainer" containerID="959f50719ef40add2c08eaa399cd8fe1c83cb7fd69ec244b942007c83407ffc8" Sep 30 14:50:51 crc kubenswrapper[4840]: E0930 14:50:51.911779 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"959f50719ef40add2c08eaa399cd8fe1c83cb7fd69ec244b942007c83407ffc8\": container with ID starting with 959f50719ef40add2c08eaa399cd8fe1c83cb7fd69ec244b942007c83407ffc8 not found: ID does not exist" containerID="959f50719ef40add2c08eaa399cd8fe1c83cb7fd69ec244b942007c83407ffc8" Sep 30 14:50:51 crc kubenswrapper[4840]: I0930 14:50:51.911815 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"959f50719ef40add2c08eaa399cd8fe1c83cb7fd69ec244b942007c83407ffc8"} err="failed to get container status \"959f50719ef40add2c08eaa399cd8fe1c83cb7fd69ec244b942007c83407ffc8\": rpc error: code = NotFound desc = could not find container \"959f50719ef40add2c08eaa399cd8fe1c83cb7fd69ec244b942007c83407ffc8\": container with ID starting with 959f50719ef40add2c08eaa399cd8fe1c83cb7fd69ec244b942007c83407ffc8 not found: ID does not exist" Sep 30 14:50:51 crc kubenswrapper[4840]: I0930 14:50:51.911840 4840 scope.go:117] "RemoveContainer" containerID="29d010a97caee4897176d4184ebd06c9389a114774136117a85bdadce370bfd4" Sep 30 14:50:51 crc kubenswrapper[4840]: E0930 14:50:51.912101 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29d010a97caee4897176d4184ebd06c9389a114774136117a85bdadce370bfd4\": container with ID starting with 29d010a97caee4897176d4184ebd06c9389a114774136117a85bdadce370bfd4 not found: ID does not exist" containerID="29d010a97caee4897176d4184ebd06c9389a114774136117a85bdadce370bfd4" Sep 30 14:50:51 crc kubenswrapper[4840]: I0930 14:50:51.912130 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29d010a97caee4897176d4184ebd06c9389a114774136117a85bdadce370bfd4"} err="failed to get container status \"29d010a97caee4897176d4184ebd06c9389a114774136117a85bdadce370bfd4\": rpc error: code = NotFound desc = could not find container \"29d010a97caee4897176d4184ebd06c9389a114774136117a85bdadce370bfd4\": container with ID starting with 29d010a97caee4897176d4184ebd06c9389a114774136117a85bdadce370bfd4 not found: ID does not exist" Sep 30 14:50:51 crc kubenswrapper[4840]: I0930 14:50:51.912150 4840 scope.go:117] "RemoveContainer" containerID="94f28a3e0f87efd2145fe14edfeaeb4a285086eb057b2c45c844a5c4d70a0d64" Sep 30 14:50:51 crc kubenswrapper[4840]: E0930 14:50:51.912356 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94f28a3e0f87efd2145fe14edfeaeb4a285086eb057b2c45c844a5c4d70a0d64\": container with ID starting with 94f28a3e0f87efd2145fe14edfeaeb4a285086eb057b2c45c844a5c4d70a0d64 not found: ID does not exist" containerID="94f28a3e0f87efd2145fe14edfeaeb4a285086eb057b2c45c844a5c4d70a0d64" Sep 30 14:50:51 crc kubenswrapper[4840]: I0930 14:50:51.912375 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94f28a3e0f87efd2145fe14edfeaeb4a285086eb057b2c45c844a5c4d70a0d64"} err="failed to get container status \"94f28a3e0f87efd2145fe14edfeaeb4a285086eb057b2c45c844a5c4d70a0d64\": rpc error: code = NotFound desc = could not find container \"94f28a3e0f87efd2145fe14edfeaeb4a285086eb057b2c45c844a5c4d70a0d64\": container with ID starting with 94f28a3e0f87efd2145fe14edfeaeb4a285086eb057b2c45c844a5c4d70a0d64 not found: ID does not exist" Sep 30 14:50:52 crc kubenswrapper[4840]: E0930 14:50:52.029613 4840 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6c84799c_2adb_492a_a005_509b4da94289.slice/crio-88fb2b9e6b9c11a9e92f2f5bf9dee6690d305fb93ac71f3a4eaf22643ec38f49\": RecentStats: unable to find data in memory cache]" Sep 30 14:50:52 crc kubenswrapper[4840]: I0930 14:50:52.128240 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c84799c-2adb-492a-a005-509b4da94289" path="/var/lib/kubelet/pods/6c84799c-2adb-492a-a005-509b4da94289/volumes" Sep 30 14:50:55 crc kubenswrapper[4840]: I0930 14:50:55.116991 4840 scope.go:117] "RemoveContainer" containerID="bb42e5820a8aaddbc79e0404192d904aa52fd67a0816b292b3605d7cf3a9d5b0" Sep 30 14:50:55 crc kubenswrapper[4840]: E0930 14:50:55.117525 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:51:07 crc kubenswrapper[4840]: I0930 14:51:07.117143 4840 scope.go:117] "RemoveContainer" containerID="bb42e5820a8aaddbc79e0404192d904aa52fd67a0816b292b3605d7cf3a9d5b0" Sep 30 14:51:07 crc kubenswrapper[4840]: E0930 14:51:07.117862 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:51:19 crc kubenswrapper[4840]: I0930 14:51:19.116480 4840 scope.go:117] "RemoveContainer" containerID="bb42e5820a8aaddbc79e0404192d904aa52fd67a0816b292b3605d7cf3a9d5b0" Sep 30 14:51:19 crc kubenswrapper[4840]: E0930 14:51:19.117276 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:51:30 crc kubenswrapper[4840]: I0930 14:51:30.121113 4840 scope.go:117] "RemoveContainer" containerID="bb42e5820a8aaddbc79e0404192d904aa52fd67a0816b292b3605d7cf3a9d5b0" Sep 30 14:51:30 crc kubenswrapper[4840]: E0930 14:51:30.121780 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:51:42 crc kubenswrapper[4840]: I0930 14:51:42.123237 4840 scope.go:117] "RemoveContainer" containerID="bb42e5820a8aaddbc79e0404192d904aa52fd67a0816b292b3605d7cf3a9d5b0" Sep 30 14:51:42 crc kubenswrapper[4840]: E0930 14:51:42.123982 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:51:54 crc kubenswrapper[4840]: I0930 14:51:54.116968 4840 scope.go:117] "RemoveContainer" containerID="bb42e5820a8aaddbc79e0404192d904aa52fd67a0816b292b3605d7cf3a9d5b0" Sep 30 14:51:54 crc kubenswrapper[4840]: E0930 14:51:54.117901 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:52:06 crc kubenswrapper[4840]: I0930 14:52:06.116222 4840 scope.go:117] "RemoveContainer" containerID="bb42e5820a8aaddbc79e0404192d904aa52fd67a0816b292b3605d7cf3a9d5b0" Sep 30 14:52:06 crc kubenswrapper[4840]: E0930 14:52:06.117036 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:52:17 crc kubenswrapper[4840]: I0930 14:52:17.116594 4840 scope.go:117] "RemoveContainer" containerID="bb42e5820a8aaddbc79e0404192d904aa52fd67a0816b292b3605d7cf3a9d5b0" Sep 30 14:52:17 crc kubenswrapper[4840]: E0930 14:52:17.117330 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:52:20 crc kubenswrapper[4840]: I0930 14:52:20.607079 4840 generic.go:334] "Generic (PLEG): container finished" podID="1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545" containerID="974d10e15595c7d52e8b9180969026db815eba4e38f1c22f60f533ae5e0f056a" exitCode=0 Sep 30 14:52:20 crc kubenswrapper[4840]: I0930 14:52:20.607180 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545","Type":"ContainerDied","Data":"974d10e15595c7d52e8b9180969026db815eba4e38f1c22f60f533ae5e0f056a"} Sep 30 14:52:21 crc kubenswrapper[4840]: I0930 14:52:21.966847 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Sep 30 14:52:22 crc kubenswrapper[4840]: I0930 14:52:22.058204 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-ssh-key\") pod \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\" (UID: \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\") " Sep 30 14:52:22 crc kubenswrapper[4840]: I0930 14:52:22.058266 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-config-data\") pod \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\" (UID: \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\") " Sep 30 14:52:22 crc kubenswrapper[4840]: I0930 14:52:22.058291 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\" (UID: \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\") " Sep 30 14:52:22 crc kubenswrapper[4840]: I0930 14:52:22.058356 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-openstack-config\") pod \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\" (UID: \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\") " Sep 30 14:52:22 crc kubenswrapper[4840]: I0930 14:52:22.058458 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-ca-certs\") pod \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\" (UID: \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\") " Sep 30 14:52:22 crc kubenswrapper[4840]: I0930 14:52:22.058618 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-test-operator-ephemeral-workdir\") pod \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\" (UID: \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\") " Sep 30 14:52:22 crc kubenswrapper[4840]: I0930 14:52:22.058661 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-openstack-config-secret\") pod \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\" (UID: \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\") " Sep 30 14:52:22 crc kubenswrapper[4840]: I0930 14:52:22.058695 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hnfwb\" (UniqueName: \"kubernetes.io/projected/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-kube-api-access-hnfwb\") pod \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\" (UID: \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\") " Sep 30 14:52:22 crc kubenswrapper[4840]: I0930 14:52:22.058746 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-test-operator-ephemeral-temporary\") pod \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\" (UID: \"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545\") " Sep 30 14:52:22 crc kubenswrapper[4840]: I0930 14:52:22.059908 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545" (UID: "1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:52:22 crc kubenswrapper[4840]: I0930 14:52:22.065180 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-config-data" (OuterVolumeSpecName: "config-data") pod "1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545" (UID: "1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:52:22 crc kubenswrapper[4840]: I0930 14:52:22.083001 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545" (UID: "1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:52:22 crc kubenswrapper[4840]: I0930 14:52:22.086419 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-kube-api-access-hnfwb" (OuterVolumeSpecName: "kube-api-access-hnfwb") pod "1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545" (UID: "1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545"). InnerVolumeSpecName "kube-api-access-hnfwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:52:22 crc kubenswrapper[4840]: I0930 14:52:22.113722 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "test-operator-logs") pod "1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545" (UID: "1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 30 14:52:22 crc kubenswrapper[4840]: I0930 14:52:22.168264 4840 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Sep 30 14:52:22 crc kubenswrapper[4840]: I0930 14:52:22.168427 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hnfwb\" (UniqueName: \"kubernetes.io/projected/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-kube-api-access-hnfwb\") on node \"crc\" DevicePath \"\"" Sep 30 14:52:22 crc kubenswrapper[4840]: I0930 14:52:22.168441 4840 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Sep 30 14:52:22 crc kubenswrapper[4840]: I0930 14:52:22.168453 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:52:22 crc kubenswrapper[4840]: I0930 14:52:22.168480 4840 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Sep 30 14:52:22 crc kubenswrapper[4840]: I0930 14:52:22.169708 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545" (UID: "1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:52:22 crc kubenswrapper[4840]: I0930 14:52:22.196779 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545" (UID: "1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:52:22 crc kubenswrapper[4840]: I0930 14:52:22.198740 4840 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Sep 30 14:52:22 crc kubenswrapper[4840]: I0930 14:52:22.211845 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545" (UID: "1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:52:22 crc kubenswrapper[4840]: I0930 14:52:22.212126 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545" (UID: "1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:52:22 crc kubenswrapper[4840]: I0930 14:52:22.270835 4840 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-ca-certs\") on node \"crc\" DevicePath \"\"" Sep 30 14:52:22 crc kubenswrapper[4840]: I0930 14:52:22.271056 4840 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Sep 30 14:52:22 crc kubenswrapper[4840]: I0930 14:52:22.271068 4840 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:52:22 crc kubenswrapper[4840]: I0930 14:52:22.271078 4840 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Sep 30 14:52:22 crc kubenswrapper[4840]: I0930 14:52:22.271088 4840 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545-openstack-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:52:22 crc kubenswrapper[4840]: I0930 14:52:22.624937 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545","Type":"ContainerDied","Data":"a4bae75b2bc474b3ecd0feba2bfdebbc43888b95aa8e1cef74f966c86bc5c243"} Sep 30 14:52:22 crc kubenswrapper[4840]: I0930 14:52:22.624988 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4bae75b2bc474b3ecd0feba2bfdebbc43888b95aa8e1cef74f966c86bc5c243" Sep 30 14:52:22 crc kubenswrapper[4840]: I0930 14:52:22.624997 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Sep 30 14:52:28 crc kubenswrapper[4840]: I0930 14:52:28.321846 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Sep 30 14:52:28 crc kubenswrapper[4840]: E0930 14:52:28.322829 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c84799c-2adb-492a-a005-509b4da94289" containerName="extract-utilities" Sep 30 14:52:28 crc kubenswrapper[4840]: I0930 14:52:28.322845 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c84799c-2adb-492a-a005-509b4da94289" containerName="extract-utilities" Sep 30 14:52:28 crc kubenswrapper[4840]: E0930 14:52:28.322863 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545" containerName="tempest-tests-tempest-tests-runner" Sep 30 14:52:28 crc kubenswrapper[4840]: I0930 14:52:28.322869 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545" containerName="tempest-tests-tempest-tests-runner" Sep 30 14:52:28 crc kubenswrapper[4840]: E0930 14:52:28.322896 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c84799c-2adb-492a-a005-509b4da94289" containerName="extract-content" Sep 30 14:52:28 crc kubenswrapper[4840]: I0930 14:52:28.322904 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c84799c-2adb-492a-a005-509b4da94289" containerName="extract-content" Sep 30 14:52:28 crc kubenswrapper[4840]: E0930 14:52:28.322920 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c84799c-2adb-492a-a005-509b4da94289" containerName="registry-server" Sep 30 14:52:28 crc kubenswrapper[4840]: I0930 14:52:28.322927 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c84799c-2adb-492a-a005-509b4da94289" containerName="registry-server" Sep 30 14:52:28 crc kubenswrapper[4840]: I0930 14:52:28.323099 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c84799c-2adb-492a-a005-509b4da94289" containerName="registry-server" Sep 30 14:52:28 crc kubenswrapper[4840]: I0930 14:52:28.323133 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545" containerName="tempest-tests-tempest-tests-runner" Sep 30 14:52:28 crc kubenswrapper[4840]: I0930 14:52:28.323835 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Sep 30 14:52:28 crc kubenswrapper[4840]: I0930 14:52:28.327095 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-vfxhm" Sep 30 14:52:28 crc kubenswrapper[4840]: I0930 14:52:28.331291 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Sep 30 14:52:28 crc kubenswrapper[4840]: I0930 14:52:28.388050 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bw7l\" (UniqueName: \"kubernetes.io/projected/caaa008d-dbdc-499e-9b2a-95c99517e733-kube-api-access-7bw7l\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"caaa008d-dbdc-499e-9b2a-95c99517e733\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Sep 30 14:52:28 crc kubenswrapper[4840]: I0930 14:52:28.388197 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"caaa008d-dbdc-499e-9b2a-95c99517e733\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Sep 30 14:52:28 crc kubenswrapper[4840]: I0930 14:52:28.490020 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bw7l\" (UniqueName: \"kubernetes.io/projected/caaa008d-dbdc-499e-9b2a-95c99517e733-kube-api-access-7bw7l\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"caaa008d-dbdc-499e-9b2a-95c99517e733\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Sep 30 14:52:28 crc kubenswrapper[4840]: I0930 14:52:28.490119 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"caaa008d-dbdc-499e-9b2a-95c99517e733\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Sep 30 14:52:28 crc kubenswrapper[4840]: I0930 14:52:28.490854 4840 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"caaa008d-dbdc-499e-9b2a-95c99517e733\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Sep 30 14:52:28 crc kubenswrapper[4840]: I0930 14:52:28.526465 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bw7l\" (UniqueName: \"kubernetes.io/projected/caaa008d-dbdc-499e-9b2a-95c99517e733-kube-api-access-7bw7l\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"caaa008d-dbdc-499e-9b2a-95c99517e733\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Sep 30 14:52:28 crc kubenswrapper[4840]: I0930 14:52:28.537424 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"caaa008d-dbdc-499e-9b2a-95c99517e733\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Sep 30 14:52:28 crc kubenswrapper[4840]: I0930 14:52:28.644358 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Sep 30 14:52:29 crc kubenswrapper[4840]: I0930 14:52:29.101414 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Sep 30 14:52:29 crc kubenswrapper[4840]: I0930 14:52:29.685401 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"caaa008d-dbdc-499e-9b2a-95c99517e733","Type":"ContainerStarted","Data":"c8e102f1d232165fb3453485823535dc69033aa3ca8c8a596eb9dab43503da04"} Sep 30 14:52:30 crc kubenswrapper[4840]: I0930 14:52:30.118244 4840 scope.go:117] "RemoveContainer" containerID="bb42e5820a8aaddbc79e0404192d904aa52fd67a0816b292b3605d7cf3a9d5b0" Sep 30 14:52:30 crc kubenswrapper[4840]: E0930 14:52:30.118862 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:52:30 crc kubenswrapper[4840]: I0930 14:52:30.695194 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"caaa008d-dbdc-499e-9b2a-95c99517e733","Type":"ContainerStarted","Data":"2e700d1094792c403ade304c5cca6cc1e3c5b3e0c6517a8e99f8f36c183dd49f"} Sep 30 14:52:30 crc kubenswrapper[4840]: I0930 14:52:30.711264 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.601286282 podStartE2EDuration="2.711239972s" podCreationTimestamp="2025-09-30 14:52:28 +0000 UTC" firstStartedPulling="2025-09-30 14:52:29.097732818 +0000 UTC m=+3377.726819241" lastFinishedPulling="2025-09-30 14:52:30.207686508 +0000 UTC m=+3378.836772931" observedRunningTime="2025-09-30 14:52:30.708604124 +0000 UTC m=+3379.337690557" watchObservedRunningTime="2025-09-30 14:52:30.711239972 +0000 UTC m=+3379.340326385" Sep 30 14:52:43 crc kubenswrapper[4840]: I0930 14:52:43.116624 4840 scope.go:117] "RemoveContainer" containerID="bb42e5820a8aaddbc79e0404192d904aa52fd67a0816b292b3605d7cf3a9d5b0" Sep 30 14:52:43 crc kubenswrapper[4840]: E0930 14:52:43.117523 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:52:48 crc kubenswrapper[4840]: I0930 14:52:48.864939 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-r4p2h/must-gather-hmmcl"] Sep 30 14:52:48 crc kubenswrapper[4840]: I0930 14:52:48.866962 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-r4p2h/must-gather-hmmcl" Sep 30 14:52:48 crc kubenswrapper[4840]: I0930 14:52:48.869186 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-r4p2h"/"kube-root-ca.crt" Sep 30 14:52:48 crc kubenswrapper[4840]: I0930 14:52:48.873020 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-r4p2h"/"openshift-service-ca.crt" Sep 30 14:52:48 crc kubenswrapper[4840]: I0930 14:52:48.914011 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-r4p2h/must-gather-hmmcl"] Sep 30 14:52:48 crc kubenswrapper[4840]: I0930 14:52:48.976543 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wbrx\" (UniqueName: \"kubernetes.io/projected/bcfd9cee-e287-458b-a32a-4f4cf71a2390-kube-api-access-7wbrx\") pod \"must-gather-hmmcl\" (UID: \"bcfd9cee-e287-458b-a32a-4f4cf71a2390\") " pod="openshift-must-gather-r4p2h/must-gather-hmmcl" Sep 30 14:52:48 crc kubenswrapper[4840]: I0930 14:52:48.976659 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bcfd9cee-e287-458b-a32a-4f4cf71a2390-must-gather-output\") pod \"must-gather-hmmcl\" (UID: \"bcfd9cee-e287-458b-a32a-4f4cf71a2390\") " pod="openshift-must-gather-r4p2h/must-gather-hmmcl" Sep 30 14:52:49 crc kubenswrapper[4840]: I0930 14:52:49.078377 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wbrx\" (UniqueName: \"kubernetes.io/projected/bcfd9cee-e287-458b-a32a-4f4cf71a2390-kube-api-access-7wbrx\") pod \"must-gather-hmmcl\" (UID: \"bcfd9cee-e287-458b-a32a-4f4cf71a2390\") " pod="openshift-must-gather-r4p2h/must-gather-hmmcl" Sep 30 14:52:49 crc kubenswrapper[4840]: I0930 14:52:49.078704 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bcfd9cee-e287-458b-a32a-4f4cf71a2390-must-gather-output\") pod \"must-gather-hmmcl\" (UID: \"bcfd9cee-e287-458b-a32a-4f4cf71a2390\") " pod="openshift-must-gather-r4p2h/must-gather-hmmcl" Sep 30 14:52:49 crc kubenswrapper[4840]: I0930 14:52:49.079057 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bcfd9cee-e287-458b-a32a-4f4cf71a2390-must-gather-output\") pod \"must-gather-hmmcl\" (UID: \"bcfd9cee-e287-458b-a32a-4f4cf71a2390\") " pod="openshift-must-gather-r4p2h/must-gather-hmmcl" Sep 30 14:52:49 crc kubenswrapper[4840]: I0930 14:52:49.101966 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wbrx\" (UniqueName: \"kubernetes.io/projected/bcfd9cee-e287-458b-a32a-4f4cf71a2390-kube-api-access-7wbrx\") pod \"must-gather-hmmcl\" (UID: \"bcfd9cee-e287-458b-a32a-4f4cf71a2390\") " pod="openshift-must-gather-r4p2h/must-gather-hmmcl" Sep 30 14:52:49 crc kubenswrapper[4840]: I0930 14:52:49.187109 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-r4p2h/must-gather-hmmcl" Sep 30 14:52:49 crc kubenswrapper[4840]: I0930 14:52:49.607147 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-r4p2h/must-gather-hmmcl"] Sep 30 14:52:49 crc kubenswrapper[4840]: W0930 14:52:49.607847 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbcfd9cee_e287_458b_a32a_4f4cf71a2390.slice/crio-6f32f3a12782264568786f135dadad4d4a97464a6d41474d2608d0500ba19aaf WatchSource:0}: Error finding container 6f32f3a12782264568786f135dadad4d4a97464a6d41474d2608d0500ba19aaf: Status 404 returned error can't find the container with id 6f32f3a12782264568786f135dadad4d4a97464a6d41474d2608d0500ba19aaf Sep 30 14:52:49 crc kubenswrapper[4840]: I0930 14:52:49.610088 4840 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 14:52:49 crc kubenswrapper[4840]: I0930 14:52:49.869728 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-r4p2h/must-gather-hmmcl" event={"ID":"bcfd9cee-e287-458b-a32a-4f4cf71a2390","Type":"ContainerStarted","Data":"6f32f3a12782264568786f135dadad4d4a97464a6d41474d2608d0500ba19aaf"} Sep 30 14:52:53 crc kubenswrapper[4840]: I0930 14:52:53.910309 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-r4p2h/must-gather-hmmcl" event={"ID":"bcfd9cee-e287-458b-a32a-4f4cf71a2390","Type":"ContainerStarted","Data":"a881c4baf61c994d460d0ca76822675954ec1bc32aabdc5086385562a9d8b88c"} Sep 30 14:52:53 crc kubenswrapper[4840]: I0930 14:52:53.910945 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-r4p2h/must-gather-hmmcl" event={"ID":"bcfd9cee-e287-458b-a32a-4f4cf71a2390","Type":"ContainerStarted","Data":"d651f5a0f9580049b759448f8c688a0a6c815f91878f45e22d1ac7a41b9badc2"} Sep 30 14:52:53 crc kubenswrapper[4840]: I0930 14:52:53.929902 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-r4p2h/must-gather-hmmcl" podStartSLOduration=2.297702964 podStartE2EDuration="5.929881451s" podCreationTimestamp="2025-09-30 14:52:48 +0000 UTC" firstStartedPulling="2025-09-30 14:52:49.610037948 +0000 UTC m=+3398.239124371" lastFinishedPulling="2025-09-30 14:52:53.242216435 +0000 UTC m=+3401.871302858" observedRunningTime="2025-09-30 14:52:53.927822258 +0000 UTC m=+3402.556908681" watchObservedRunningTime="2025-09-30 14:52:53.929881451 +0000 UTC m=+3402.558967874" Sep 30 14:52:56 crc kubenswrapper[4840]: I0930 14:52:56.705882 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-r4p2h/crc-debug-7zq9l"] Sep 30 14:52:56 crc kubenswrapper[4840]: I0930 14:52:56.707275 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-r4p2h/crc-debug-7zq9l" Sep 30 14:52:56 crc kubenswrapper[4840]: I0930 14:52:56.711230 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-r4p2h"/"default-dockercfg-pv7db" Sep 30 14:52:56 crc kubenswrapper[4840]: I0930 14:52:56.825892 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbvvm\" (UniqueName: \"kubernetes.io/projected/bf9ca252-8b08-45b5-bd98-442bece17347-kube-api-access-gbvvm\") pod \"crc-debug-7zq9l\" (UID: \"bf9ca252-8b08-45b5-bd98-442bece17347\") " pod="openshift-must-gather-r4p2h/crc-debug-7zq9l" Sep 30 14:52:56 crc kubenswrapper[4840]: I0930 14:52:56.825942 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bf9ca252-8b08-45b5-bd98-442bece17347-host\") pod \"crc-debug-7zq9l\" (UID: \"bf9ca252-8b08-45b5-bd98-442bece17347\") " pod="openshift-must-gather-r4p2h/crc-debug-7zq9l" Sep 30 14:52:56 crc kubenswrapper[4840]: I0930 14:52:56.927683 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbvvm\" (UniqueName: \"kubernetes.io/projected/bf9ca252-8b08-45b5-bd98-442bece17347-kube-api-access-gbvvm\") pod \"crc-debug-7zq9l\" (UID: \"bf9ca252-8b08-45b5-bd98-442bece17347\") " pod="openshift-must-gather-r4p2h/crc-debug-7zq9l" Sep 30 14:52:56 crc kubenswrapper[4840]: I0930 14:52:56.927742 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bf9ca252-8b08-45b5-bd98-442bece17347-host\") pod \"crc-debug-7zq9l\" (UID: \"bf9ca252-8b08-45b5-bd98-442bece17347\") " pod="openshift-must-gather-r4p2h/crc-debug-7zq9l" Sep 30 14:52:56 crc kubenswrapper[4840]: I0930 14:52:56.927850 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bf9ca252-8b08-45b5-bd98-442bece17347-host\") pod \"crc-debug-7zq9l\" (UID: \"bf9ca252-8b08-45b5-bd98-442bece17347\") " pod="openshift-must-gather-r4p2h/crc-debug-7zq9l" Sep 30 14:52:56 crc kubenswrapper[4840]: I0930 14:52:56.948072 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbvvm\" (UniqueName: \"kubernetes.io/projected/bf9ca252-8b08-45b5-bd98-442bece17347-kube-api-access-gbvvm\") pod \"crc-debug-7zq9l\" (UID: \"bf9ca252-8b08-45b5-bd98-442bece17347\") " pod="openshift-must-gather-r4p2h/crc-debug-7zq9l" Sep 30 14:52:57 crc kubenswrapper[4840]: I0930 14:52:57.025726 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-r4p2h/crc-debug-7zq9l" Sep 30 14:52:57 crc kubenswrapper[4840]: I0930 14:52:57.950456 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-r4p2h/crc-debug-7zq9l" event={"ID":"bf9ca252-8b08-45b5-bd98-442bece17347","Type":"ContainerStarted","Data":"efd9ee9be2374911ba81c805899938fccbfe4c1e4ae0309324795953a2444d19"} Sep 30 14:52:58 crc kubenswrapper[4840]: I0930 14:52:58.116845 4840 scope.go:117] "RemoveContainer" containerID="bb42e5820a8aaddbc79e0404192d904aa52fd67a0816b292b3605d7cf3a9d5b0" Sep 30 14:52:58 crc kubenswrapper[4840]: E0930 14:52:58.117109 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:53:08 crc kubenswrapper[4840]: I0930 14:53:08.051032 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-r4p2h/crc-debug-7zq9l" event={"ID":"bf9ca252-8b08-45b5-bd98-442bece17347","Type":"ContainerStarted","Data":"080b4a392d725b13e94a18caad6a5fb0fd3d342d758f311ad9ec204fa2b1a720"} Sep 30 14:53:08 crc kubenswrapper[4840]: I0930 14:53:08.067202 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-r4p2h/crc-debug-7zq9l" podStartSLOduration=2.265704697 podStartE2EDuration="12.067185542s" podCreationTimestamp="2025-09-30 14:52:56 +0000 UTC" firstStartedPulling="2025-09-30 14:52:57.063060691 +0000 UTC m=+3405.692147104" lastFinishedPulling="2025-09-30 14:53:06.864541516 +0000 UTC m=+3415.493627949" observedRunningTime="2025-09-30 14:53:08.065232092 +0000 UTC m=+3416.694318505" watchObservedRunningTime="2025-09-30 14:53:08.067185542 +0000 UTC m=+3416.696271965" Sep 30 14:53:13 crc kubenswrapper[4840]: I0930 14:53:13.116418 4840 scope.go:117] "RemoveContainer" containerID="bb42e5820a8aaddbc79e0404192d904aa52fd67a0816b292b3605d7cf3a9d5b0" Sep 30 14:53:13 crc kubenswrapper[4840]: E0930 14:53:13.117334 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 14:53:24 crc kubenswrapper[4840]: I0930 14:53:24.116766 4840 scope.go:117] "RemoveContainer" containerID="bb42e5820a8aaddbc79e0404192d904aa52fd67a0816b292b3605d7cf3a9d5b0" Sep 30 14:53:25 crc kubenswrapper[4840]: I0930 14:53:25.210405 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" event={"ID":"10e8b890-7f20-4a36-8e03-898620cf599a","Type":"ContainerStarted","Data":"f4b62e43937e756d3df8cd25fd431ebd8934b3c7b61115065214e7963ec22cba"} Sep 30 14:53:53 crc kubenswrapper[4840]: I0930 14:53:53.836044 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-784f485bd4-gztrl_f7f068af-286b-4d75-ace9-247f0432de8a/barbican-api/0.log" Sep 30 14:53:53 crc kubenswrapper[4840]: I0930 14:53:53.904907 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-784f485bd4-gztrl_f7f068af-286b-4d75-ace9-247f0432de8a/barbican-api-log/0.log" Sep 30 14:53:54 crc kubenswrapper[4840]: I0930 14:53:54.087240 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-67bf9f4b94-nk9h8_ae9a65ac-f1a7-4ef0-aeba-aca57985ae87/barbican-keystone-listener/0.log" Sep 30 14:53:54 crc kubenswrapper[4840]: I0930 14:53:54.111793 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-67bf9f4b94-nk9h8_ae9a65ac-f1a7-4ef0-aeba-aca57985ae87/barbican-keystone-listener-log/0.log" Sep 30 14:53:54 crc kubenswrapper[4840]: I0930 14:53:54.283509 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-668f8fbd45-fpqbx_0fa943f1-8698-41e3-a893-fd1c8845076a/barbican-worker-log/0.log" Sep 30 14:53:54 crc kubenswrapper[4840]: I0930 14:53:54.300983 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-668f8fbd45-fpqbx_0fa943f1-8698-41e3-a893-fd1c8845076a/barbican-worker/0.log" Sep 30 14:53:54 crc kubenswrapper[4840]: I0930 14:53:54.509438 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-6ntwv_558c75bb-7947-423d-b3aa-1b2ab4ba2d1d/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 14:53:54 crc kubenswrapper[4840]: I0930 14:53:54.686991 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_5673b7b1-d1e1-47ac-8a65-c210430ac4a5/ceilometer-central-agent/0.log" Sep 30 14:53:54 crc kubenswrapper[4840]: I0930 14:53:54.708491 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_5673b7b1-d1e1-47ac-8a65-c210430ac4a5/ceilometer-notification-agent/0.log" Sep 30 14:53:54 crc kubenswrapper[4840]: I0930 14:53:54.752908 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_5673b7b1-d1e1-47ac-8a65-c210430ac4a5/proxy-httpd/0.log" Sep 30 14:53:54 crc kubenswrapper[4840]: I0930 14:53:54.873725 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_5673b7b1-d1e1-47ac-8a65-c210430ac4a5/sg-core/0.log" Sep 30 14:53:54 crc kubenswrapper[4840]: I0930 14:53:54.976012 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_eb01e3a9-61de-4dc8-954b-221d80a675fb/cinder-api/0.log" Sep 30 14:53:55 crc kubenswrapper[4840]: I0930 14:53:55.056300 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_eb01e3a9-61de-4dc8-954b-221d80a675fb/cinder-api-log/0.log" Sep 30 14:53:55 crc kubenswrapper[4840]: I0930 14:53:55.178150 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_01feefa5-d23e-4189-9d57-84bae2e247e2/cinder-scheduler/0.log" Sep 30 14:53:55 crc kubenswrapper[4840]: I0930 14:53:55.258995 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_01feefa5-d23e-4189-9d57-84bae2e247e2/probe/0.log" Sep 30 14:53:55 crc kubenswrapper[4840]: I0930 14:53:55.438730 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-xvhf8_b3a0bf4c-6d14-49c4-9a98-3506318ee798/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 14:53:55 crc kubenswrapper[4840]: I0930 14:53:55.651718 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-9p2lj_cbc06ec4-8aae-4192-a201-d2b0d13b29b1/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 14:53:55 crc kubenswrapper[4840]: I0930 14:53:55.725489 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-bmz4g_797b0ada-6af8-4509-8a81-6b97c263fdc3/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 14:53:55 crc kubenswrapper[4840]: I0930 14:53:55.886122 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-87k2l_462dd9f7-ec69-4883-aa78-9b014c449baa/init/0.log" Sep 30 14:53:56 crc kubenswrapper[4840]: I0930 14:53:56.126906 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-87k2l_462dd9f7-ec69-4883-aa78-9b014c449baa/dnsmasq-dns/0.log" Sep 30 14:53:56 crc kubenswrapper[4840]: I0930 14:53:56.227644 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-87k2l_462dd9f7-ec69-4883-aa78-9b014c449baa/init/0.log" Sep 30 14:53:56 crc kubenswrapper[4840]: I0930 14:53:56.343162 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-cbh6n_17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 14:53:56 crc kubenswrapper[4840]: I0930 14:53:56.543283 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_b1e73274-948e-4813-946c-0d98ce327e91/glance-log/0.log" Sep 30 14:53:56 crc kubenswrapper[4840]: I0930 14:53:56.551840 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_b1e73274-948e-4813-946c-0d98ce327e91/glance-httpd/0.log" Sep 30 14:53:56 crc kubenswrapper[4840]: I0930 14:53:56.848238 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_c58eeb6f-560e-4abf-b604-65a3d2819ac5/glance-httpd/0.log" Sep 30 14:53:56 crc kubenswrapper[4840]: I0930 14:53:56.889651 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_c58eeb6f-560e-4abf-b604-65a3d2819ac5/glance-log/0.log" Sep 30 14:53:57 crc kubenswrapper[4840]: I0930 14:53:57.161786 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-94dfcc48d-mlrjd_1138b51b-edf9-42d2-8ea4-4d4b9122eae4/horizon/0.log" Sep 30 14:53:57 crc kubenswrapper[4840]: I0930 14:53:57.246263 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx_cd627175-3b3b-4a69-9f1c-f64828267f40/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 14:53:57 crc kubenswrapper[4840]: I0930 14:53:57.413070 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-94dfcc48d-mlrjd_1138b51b-edf9-42d2-8ea4-4d4b9122eae4/horizon-log/0.log" Sep 30 14:53:57 crc kubenswrapper[4840]: I0930 14:53:57.454275 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-hw58j_3c378bd2-1562-466e-aafb-7a597487f2c0/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 14:53:57 crc kubenswrapper[4840]: I0930 14:53:57.659058 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_26d4f9e8-9d18-45ec-8117-39c0714ce1b0/kube-state-metrics/0.log" Sep 30 14:53:57 crc kubenswrapper[4840]: I0930 14:53:57.746878 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6c78d49dc8-kqrxd_692fd2c9-5168-48fb-9909-7b1cab678b75/keystone-api/0.log" Sep 30 14:53:57 crc kubenswrapper[4840]: I0930 14:53:57.874168 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-m584b_9f66833a-f0de-410b-9459-80203d4291a6/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 14:53:58 crc kubenswrapper[4840]: I0930 14:53:58.276175 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-658bb978b7-lrh6p_e8e55cd0-dac0-402d-beb7-6482894554fb/neutron-httpd/0.log" Sep 30 14:53:58 crc kubenswrapper[4840]: I0930 14:53:58.320417 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-658bb978b7-lrh6p_e8e55cd0-dac0-402d-beb7-6482894554fb/neutron-api/0.log" Sep 30 14:53:58 crc kubenswrapper[4840]: I0930 14:53:58.486028 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8_800dcd1d-c19d-4841-b3ab-81c312d8a754/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 14:53:58 crc kubenswrapper[4840]: I0930 14:53:58.978960 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_eade78f5-391f-455f-b61c-4300f8d9f53c/nova-api-log/0.log" Sep 30 14:53:59 crc kubenswrapper[4840]: I0930 14:53:59.066187 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_eade78f5-391f-455f-b61c-4300f8d9f53c/nova-api-api/0.log" Sep 30 14:53:59 crc kubenswrapper[4840]: I0930 14:53:59.166799 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_739533f4-30e5-4941-a7f3-d43a945cd8e8/nova-cell0-conductor-conductor/0.log" Sep 30 14:53:59 crc kubenswrapper[4840]: I0930 14:53:59.393328 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_b63a8a4d-0f06-4f78-9f54-9b4e81183705/nova-cell1-conductor-conductor/0.log" Sep 30 14:53:59 crc kubenswrapper[4840]: I0930 14:53:59.510965 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_194d271a-568f-4295-84f2-80cf46e1f123/nova-cell1-novncproxy-novncproxy/0.log" Sep 30 14:53:59 crc kubenswrapper[4840]: I0930 14:53:59.711011 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-5q27c_20b6236d-9397-4095-a35a-6c24222b852f/nova-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 14:53:59 crc kubenswrapper[4840]: I0930 14:53:59.950420 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_0f1f889e-0a43-436f-adfe-6bb0a8e9b909/nova-metadata-log/0.log" Sep 30 14:54:00 crc kubenswrapper[4840]: I0930 14:54:00.380439 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_ca7038c5-867f-4546-b846-7a6beae1fa1f/nova-scheduler-scheduler/0.log" Sep 30 14:54:00 crc kubenswrapper[4840]: I0930 14:54:00.437408 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_41b0d67b-4ef4-4ebb-b000-b37bbdac6b34/mysql-bootstrap/0.log" Sep 30 14:54:00 crc kubenswrapper[4840]: I0930 14:54:00.633392 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_41b0d67b-4ef4-4ebb-b000-b37bbdac6b34/mysql-bootstrap/0.log" Sep 30 14:54:00 crc kubenswrapper[4840]: I0930 14:54:00.688152 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_41b0d67b-4ef4-4ebb-b000-b37bbdac6b34/galera/0.log" Sep 30 14:54:00 crc kubenswrapper[4840]: I0930 14:54:00.930361 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f/mysql-bootstrap/0.log" Sep 30 14:54:01 crc kubenswrapper[4840]: I0930 14:54:01.142988 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f/mysql-bootstrap/0.log" Sep 30 14:54:01 crc kubenswrapper[4840]: I0930 14:54:01.192136 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f/galera/0.log" Sep 30 14:54:01 crc kubenswrapper[4840]: I0930 14:54:01.230638 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_0f1f889e-0a43-436f-adfe-6bb0a8e9b909/nova-metadata-metadata/0.log" Sep 30 14:54:01 crc kubenswrapper[4840]: I0930 14:54:01.409897 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_e234012b-317e-4c79-81a3-3991e1296058/openstackclient/0.log" Sep 30 14:54:01 crc kubenswrapper[4840]: I0930 14:54:01.578563 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-cngvz_e8a63b50-ecd5-4993-a890-8c94bc2d5e60/ovn-controller/0.log" Sep 30 14:54:01 crc kubenswrapper[4840]: I0930 14:54:01.688334 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-2bdgj_a789c20a-6871-4866-bfa2-c69efd876afb/openstack-network-exporter/0.log" Sep 30 14:54:01 crc kubenswrapper[4840]: I0930 14:54:01.920891 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fc5jg_7d1f7d84-d053-4c95-83ac-ecbff773463d/ovsdb-server-init/0.log" Sep 30 14:54:02 crc kubenswrapper[4840]: I0930 14:54:02.110996 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fc5jg_7d1f7d84-d053-4c95-83ac-ecbff773463d/ovsdb-server-init/0.log" Sep 30 14:54:02 crc kubenswrapper[4840]: I0930 14:54:02.130017 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fc5jg_7d1f7d84-d053-4c95-83ac-ecbff773463d/ovs-vswitchd/0.log" Sep 30 14:54:02 crc kubenswrapper[4840]: I0930 14:54:02.189882 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fc5jg_7d1f7d84-d053-4c95-83ac-ecbff773463d/ovsdb-server/0.log" Sep 30 14:54:02 crc kubenswrapper[4840]: I0930 14:54:02.358743 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-tsm4h_bc398435-386f-4d82-8ba5-618d3795223e/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 14:54:02 crc kubenswrapper[4840]: I0930 14:54:02.571997 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_2338a6b1-3726-4994-9325-ee16f93e28c0/openstack-network-exporter/0.log" Sep 30 14:54:02 crc kubenswrapper[4840]: I0930 14:54:02.590704 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_2338a6b1-3726-4994-9325-ee16f93e28c0/ovn-northd/0.log" Sep 30 14:54:02 crc kubenswrapper[4840]: I0930 14:54:02.823104 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_2082b33d-3b6c-449a-9bdd-63d665a3d313/openstack-network-exporter/0.log" Sep 30 14:54:02 crc kubenswrapper[4840]: I0930 14:54:02.846122 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_2082b33d-3b6c-449a-9bdd-63d665a3d313/ovsdbserver-nb/0.log" Sep 30 14:54:03 crc kubenswrapper[4840]: I0930 14:54:03.054209 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_93d37b01-5288-4d31-84db-3fd7e7840c3b/openstack-network-exporter/0.log" Sep 30 14:54:03 crc kubenswrapper[4840]: I0930 14:54:03.092009 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_93d37b01-5288-4d31-84db-3fd7e7840c3b/ovsdbserver-sb/0.log" Sep 30 14:54:03 crc kubenswrapper[4840]: I0930 14:54:03.358660 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6cbd5fdcfb-hn4pn_c1eaca45-6337-4f35-ac67-dfa0fc83e1bb/placement-api/0.log" Sep 30 14:54:03 crc kubenswrapper[4840]: I0930 14:54:03.400333 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6cbd5fdcfb-hn4pn_c1eaca45-6337-4f35-ac67-dfa0fc83e1bb/placement-log/0.log" Sep 30 14:54:03 crc kubenswrapper[4840]: I0930 14:54:03.578895 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_533b8943-b9ea-4062-9dd2-ed1b29d4f2c6/setup-container/0.log" Sep 30 14:54:03 crc kubenswrapper[4840]: I0930 14:54:03.746508 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_533b8943-b9ea-4062-9dd2-ed1b29d4f2c6/setup-container/0.log" Sep 30 14:54:03 crc kubenswrapper[4840]: I0930 14:54:03.842926 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_533b8943-b9ea-4062-9dd2-ed1b29d4f2c6/rabbitmq/0.log" Sep 30 14:54:03 crc kubenswrapper[4840]: I0930 14:54:03.957941 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_97b46089-3117-4a2a-b9db-ada9d0daca8b/setup-container/0.log" Sep 30 14:54:04 crc kubenswrapper[4840]: I0930 14:54:04.215898 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_97b46089-3117-4a2a-b9db-ada9d0daca8b/rabbitmq/0.log" Sep 30 14:54:04 crc kubenswrapper[4840]: I0930 14:54:04.233097 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_97b46089-3117-4a2a-b9db-ada9d0daca8b/setup-container/0.log" Sep 30 14:54:04 crc kubenswrapper[4840]: I0930 14:54:04.426286 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-xkx5r_86b7d072-1a56-454f-ad5f-a098a0977122/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 14:54:04 crc kubenswrapper[4840]: I0930 14:54:04.552361 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-gtxqx_84f24beb-97af-44e4-b529-41138e5414ad/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 14:54:04 crc kubenswrapper[4840]: I0930 14:54:04.699642 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-q5t88_c1073e6a-94e1-4fad-a116-ff4e145a0d22/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 14:54:04 crc kubenswrapper[4840]: I0930 14:54:04.891766 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-nn4ls_a7b69c13-1a32-4676-9912-e117b0b59848/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 14:54:05 crc kubenswrapper[4840]: I0930 14:54:05.060336 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-dbvwh_37a41ba6-72d7-40d8-a305-153b68f100a8/ssh-known-hosts-edpm-deployment/0.log" Sep 30 14:54:05 crc kubenswrapper[4840]: I0930 14:54:05.291032 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5d586fc9-9z92d_84813c70-f94c-4556-a498-10e17fb354f1/proxy-server/0.log" Sep 30 14:54:05 crc kubenswrapper[4840]: I0930 14:54:05.314093 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5d586fc9-9z92d_84813c70-f94c-4556-a498-10e17fb354f1/proxy-httpd/0.log" Sep 30 14:54:05 crc kubenswrapper[4840]: I0930 14:54:05.480700 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-txwn9_d8412105-132d-4656-8fd8-296cbd8a1ace/swift-ring-rebalance/0.log" Sep 30 14:54:05 crc kubenswrapper[4840]: I0930 14:54:05.557391 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66a7a4e6-dd49-4589-bd88-2f5fdf0af70f/account-auditor/0.log" Sep 30 14:54:05 crc kubenswrapper[4840]: I0930 14:54:05.705183 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66a7a4e6-dd49-4589-bd88-2f5fdf0af70f/account-reaper/0.log" Sep 30 14:54:05 crc kubenswrapper[4840]: I0930 14:54:05.737056 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66a7a4e6-dd49-4589-bd88-2f5fdf0af70f/account-server/0.log" Sep 30 14:54:05 crc kubenswrapper[4840]: I0930 14:54:05.848425 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66a7a4e6-dd49-4589-bd88-2f5fdf0af70f/account-replicator/0.log" Sep 30 14:54:05 crc kubenswrapper[4840]: I0930 14:54:05.922001 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66a7a4e6-dd49-4589-bd88-2f5fdf0af70f/container-auditor/0.log" Sep 30 14:54:05 crc kubenswrapper[4840]: I0930 14:54:05.965021 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66a7a4e6-dd49-4589-bd88-2f5fdf0af70f/container-replicator/0.log" Sep 30 14:54:06 crc kubenswrapper[4840]: I0930 14:54:06.082826 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66a7a4e6-dd49-4589-bd88-2f5fdf0af70f/container-server/0.log" Sep 30 14:54:06 crc kubenswrapper[4840]: I0930 14:54:06.136369 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66a7a4e6-dd49-4589-bd88-2f5fdf0af70f/container-updater/0.log" Sep 30 14:54:06 crc kubenswrapper[4840]: I0930 14:54:06.211350 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66a7a4e6-dd49-4589-bd88-2f5fdf0af70f/object-auditor/0.log" Sep 30 14:54:06 crc kubenswrapper[4840]: I0930 14:54:06.286285 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66a7a4e6-dd49-4589-bd88-2f5fdf0af70f/object-expirer/0.log" Sep 30 14:54:06 crc kubenswrapper[4840]: I0930 14:54:06.360639 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66a7a4e6-dd49-4589-bd88-2f5fdf0af70f/object-replicator/0.log" Sep 30 14:54:06 crc kubenswrapper[4840]: I0930 14:54:06.447424 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66a7a4e6-dd49-4589-bd88-2f5fdf0af70f/object-server/0.log" Sep 30 14:54:06 crc kubenswrapper[4840]: I0930 14:54:06.515679 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66a7a4e6-dd49-4589-bd88-2f5fdf0af70f/object-updater/0.log" Sep 30 14:54:06 crc kubenswrapper[4840]: I0930 14:54:06.599480 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66a7a4e6-dd49-4589-bd88-2f5fdf0af70f/rsync/0.log" Sep 30 14:54:06 crc kubenswrapper[4840]: I0930 14:54:06.694256 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66a7a4e6-dd49-4589-bd88-2f5fdf0af70f/swift-recon-cron/0.log" Sep 30 14:54:06 crc kubenswrapper[4840]: I0930 14:54:06.845053 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-dzddj_73a8aa6c-814e-420f-bb1d-666259304a7e/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 14:54:06 crc kubenswrapper[4840]: I0930 14:54:06.950710 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545/tempest-tests-tempest-tests-runner/0.log" Sep 30 14:54:07 crc kubenswrapper[4840]: I0930 14:54:07.155036 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_caaa008d-dbdc-499e-9b2a-95c99517e733/test-operator-logs-container/0.log" Sep 30 14:54:07 crc kubenswrapper[4840]: I0930 14:54:07.315747 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-wmtv5_43ad4b75-0f40-4adc-8bda-67ed967051f3/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 14:54:16 crc kubenswrapper[4840]: I0930 14:54:16.662928 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_f0207213-00d5-4b63-b041-fff191463f4a/memcached/0.log" Sep 30 14:54:37 crc kubenswrapper[4840]: I0930 14:54:37.901455 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-h4pnt"] Sep 30 14:54:37 crc kubenswrapper[4840]: I0930 14:54:37.904716 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h4pnt" Sep 30 14:54:37 crc kubenswrapper[4840]: I0930 14:54:37.923246 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h4pnt"] Sep 30 14:54:37 crc kubenswrapper[4840]: I0930 14:54:37.998860 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42zvz\" (UniqueName: \"kubernetes.io/projected/4de366b7-3778-44c5-8187-4d0bdfdbbbc1-kube-api-access-42zvz\") pod \"community-operators-h4pnt\" (UID: \"4de366b7-3778-44c5-8187-4d0bdfdbbbc1\") " pod="openshift-marketplace/community-operators-h4pnt" Sep 30 14:54:37 crc kubenswrapper[4840]: I0930 14:54:37.998926 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4de366b7-3778-44c5-8187-4d0bdfdbbbc1-catalog-content\") pod \"community-operators-h4pnt\" (UID: \"4de366b7-3778-44c5-8187-4d0bdfdbbbc1\") " pod="openshift-marketplace/community-operators-h4pnt" Sep 30 14:54:37 crc kubenswrapper[4840]: I0930 14:54:37.998979 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4de366b7-3778-44c5-8187-4d0bdfdbbbc1-utilities\") pod \"community-operators-h4pnt\" (UID: \"4de366b7-3778-44c5-8187-4d0bdfdbbbc1\") " pod="openshift-marketplace/community-operators-h4pnt" Sep 30 14:54:38 crc kubenswrapper[4840]: I0930 14:54:38.100485 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4de366b7-3778-44c5-8187-4d0bdfdbbbc1-catalog-content\") pod \"community-operators-h4pnt\" (UID: \"4de366b7-3778-44c5-8187-4d0bdfdbbbc1\") " pod="openshift-marketplace/community-operators-h4pnt" Sep 30 14:54:38 crc kubenswrapper[4840]: I0930 14:54:38.100842 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4de366b7-3778-44c5-8187-4d0bdfdbbbc1-utilities\") pod \"community-operators-h4pnt\" (UID: \"4de366b7-3778-44c5-8187-4d0bdfdbbbc1\") " pod="openshift-marketplace/community-operators-h4pnt" Sep 30 14:54:38 crc kubenswrapper[4840]: I0930 14:54:38.101133 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42zvz\" (UniqueName: \"kubernetes.io/projected/4de366b7-3778-44c5-8187-4d0bdfdbbbc1-kube-api-access-42zvz\") pod \"community-operators-h4pnt\" (UID: \"4de366b7-3778-44c5-8187-4d0bdfdbbbc1\") " pod="openshift-marketplace/community-operators-h4pnt" Sep 30 14:54:38 crc kubenswrapper[4840]: I0930 14:54:38.101534 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4de366b7-3778-44c5-8187-4d0bdfdbbbc1-utilities\") pod \"community-operators-h4pnt\" (UID: \"4de366b7-3778-44c5-8187-4d0bdfdbbbc1\") " pod="openshift-marketplace/community-operators-h4pnt" Sep 30 14:54:38 crc kubenswrapper[4840]: I0930 14:54:38.101591 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4de366b7-3778-44c5-8187-4d0bdfdbbbc1-catalog-content\") pod \"community-operators-h4pnt\" (UID: \"4de366b7-3778-44c5-8187-4d0bdfdbbbc1\") " pod="openshift-marketplace/community-operators-h4pnt" Sep 30 14:54:38 crc kubenswrapper[4840]: I0930 14:54:38.125464 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42zvz\" (UniqueName: \"kubernetes.io/projected/4de366b7-3778-44c5-8187-4d0bdfdbbbc1-kube-api-access-42zvz\") pod \"community-operators-h4pnt\" (UID: \"4de366b7-3778-44c5-8187-4d0bdfdbbbc1\") " pod="openshift-marketplace/community-operators-h4pnt" Sep 30 14:54:38 crc kubenswrapper[4840]: I0930 14:54:38.238678 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h4pnt" Sep 30 14:54:38 crc kubenswrapper[4840]: I0930 14:54:38.839159 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h4pnt"] Sep 30 14:54:38 crc kubenswrapper[4840]: I0930 14:54:38.856314 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h4pnt" event={"ID":"4de366b7-3778-44c5-8187-4d0bdfdbbbc1","Type":"ContainerStarted","Data":"cd337b980a30612958f844616e9119a00a5b670958c8f365ea95fbddd74e7f6f"} Sep 30 14:54:39 crc kubenswrapper[4840]: I0930 14:54:39.868644 4840 generic.go:334] "Generic (PLEG): container finished" podID="4de366b7-3778-44c5-8187-4d0bdfdbbbc1" containerID="b2798b768be368e63b9ab7c81588d32e4f99494753ea7a1cbc0ee53cb1ea3ab1" exitCode=0 Sep 30 14:54:39 crc kubenswrapper[4840]: I0930 14:54:39.868946 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h4pnt" event={"ID":"4de366b7-3778-44c5-8187-4d0bdfdbbbc1","Type":"ContainerDied","Data":"b2798b768be368e63b9ab7c81588d32e4f99494753ea7a1cbc0ee53cb1ea3ab1"} Sep 30 14:54:40 crc kubenswrapper[4840]: I0930 14:54:40.888957 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h4pnt" event={"ID":"4de366b7-3778-44c5-8187-4d0bdfdbbbc1","Type":"ContainerStarted","Data":"a2dfc38f8fe8ac16801edd633580c3de4ff4f7266676c5c5b8f552aec7a10e92"} Sep 30 14:54:41 crc kubenswrapper[4840]: I0930 14:54:41.899007 4840 generic.go:334] "Generic (PLEG): container finished" podID="4de366b7-3778-44c5-8187-4d0bdfdbbbc1" containerID="a2dfc38f8fe8ac16801edd633580c3de4ff4f7266676c5c5b8f552aec7a10e92" exitCode=0 Sep 30 14:54:41 crc kubenswrapper[4840]: I0930 14:54:41.899181 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h4pnt" event={"ID":"4de366b7-3778-44c5-8187-4d0bdfdbbbc1","Type":"ContainerDied","Data":"a2dfc38f8fe8ac16801edd633580c3de4ff4f7266676c5c5b8f552aec7a10e92"} Sep 30 14:54:42 crc kubenswrapper[4840]: I0930 14:54:42.908733 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h4pnt" event={"ID":"4de366b7-3778-44c5-8187-4d0bdfdbbbc1","Type":"ContainerStarted","Data":"620bf9c8d80335be61d4377df7f11539c5e4b98bebed586664a5842e861c97f8"} Sep 30 14:54:42 crc kubenswrapper[4840]: I0930 14:54:42.937966 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-h4pnt" podStartSLOduration=3.416433354 podStartE2EDuration="5.937944602s" podCreationTimestamp="2025-09-30 14:54:37 +0000 UTC" firstStartedPulling="2025-09-30 14:54:39.871434637 +0000 UTC m=+3508.500521090" lastFinishedPulling="2025-09-30 14:54:42.392945895 +0000 UTC m=+3511.022032338" observedRunningTime="2025-09-30 14:54:42.930825548 +0000 UTC m=+3511.559911971" watchObservedRunningTime="2025-09-30 14:54:42.937944602 +0000 UTC m=+3511.567031055" Sep 30 14:54:48 crc kubenswrapper[4840]: I0930 14:54:48.239152 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-h4pnt" Sep 30 14:54:48 crc kubenswrapper[4840]: I0930 14:54:48.240806 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-h4pnt" Sep 30 14:54:48 crc kubenswrapper[4840]: I0930 14:54:48.285370 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-h4pnt" Sep 30 14:54:49 crc kubenswrapper[4840]: I0930 14:54:49.019806 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-h4pnt" Sep 30 14:54:49 crc kubenswrapper[4840]: I0930 14:54:49.081466 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-h4pnt"] Sep 30 14:54:50 crc kubenswrapper[4840]: I0930 14:54:50.989753 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-h4pnt" podUID="4de366b7-3778-44c5-8187-4d0bdfdbbbc1" containerName="registry-server" containerID="cri-o://620bf9c8d80335be61d4377df7f11539c5e4b98bebed586664a5842e861c97f8" gracePeriod=2 Sep 30 14:54:51 crc kubenswrapper[4840]: I0930 14:54:51.489390 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h4pnt" Sep 30 14:54:51 crc kubenswrapper[4840]: I0930 14:54:51.570480 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4de366b7-3778-44c5-8187-4d0bdfdbbbc1-utilities\") pod \"4de366b7-3778-44c5-8187-4d0bdfdbbbc1\" (UID: \"4de366b7-3778-44c5-8187-4d0bdfdbbbc1\") " Sep 30 14:54:51 crc kubenswrapper[4840]: I0930 14:54:51.570575 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42zvz\" (UniqueName: \"kubernetes.io/projected/4de366b7-3778-44c5-8187-4d0bdfdbbbc1-kube-api-access-42zvz\") pod \"4de366b7-3778-44c5-8187-4d0bdfdbbbc1\" (UID: \"4de366b7-3778-44c5-8187-4d0bdfdbbbc1\") " Sep 30 14:54:51 crc kubenswrapper[4840]: I0930 14:54:51.571332 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4de366b7-3778-44c5-8187-4d0bdfdbbbc1-utilities" (OuterVolumeSpecName: "utilities") pod "4de366b7-3778-44c5-8187-4d0bdfdbbbc1" (UID: "4de366b7-3778-44c5-8187-4d0bdfdbbbc1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:54:51 crc kubenswrapper[4840]: I0930 14:54:51.571974 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4de366b7-3778-44c5-8187-4d0bdfdbbbc1-catalog-content\") pod \"4de366b7-3778-44c5-8187-4d0bdfdbbbc1\" (UID: \"4de366b7-3778-44c5-8187-4d0bdfdbbbc1\") " Sep 30 14:54:51 crc kubenswrapper[4840]: I0930 14:54:51.572729 4840 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4de366b7-3778-44c5-8187-4d0bdfdbbbc1-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:54:51 crc kubenswrapper[4840]: I0930 14:54:51.578745 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4de366b7-3778-44c5-8187-4d0bdfdbbbc1-kube-api-access-42zvz" (OuterVolumeSpecName: "kube-api-access-42zvz") pod "4de366b7-3778-44c5-8187-4d0bdfdbbbc1" (UID: "4de366b7-3778-44c5-8187-4d0bdfdbbbc1"). InnerVolumeSpecName "kube-api-access-42zvz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:54:51 crc kubenswrapper[4840]: I0930 14:54:51.617475 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4de366b7-3778-44c5-8187-4d0bdfdbbbc1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4de366b7-3778-44c5-8187-4d0bdfdbbbc1" (UID: "4de366b7-3778-44c5-8187-4d0bdfdbbbc1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:54:51 crc kubenswrapper[4840]: I0930 14:54:51.674638 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42zvz\" (UniqueName: \"kubernetes.io/projected/4de366b7-3778-44c5-8187-4d0bdfdbbbc1-kube-api-access-42zvz\") on node \"crc\" DevicePath \"\"" Sep 30 14:54:51 crc kubenswrapper[4840]: I0930 14:54:51.674674 4840 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4de366b7-3778-44c5-8187-4d0bdfdbbbc1-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:54:52 crc kubenswrapper[4840]: I0930 14:54:52.014257 4840 generic.go:334] "Generic (PLEG): container finished" podID="4de366b7-3778-44c5-8187-4d0bdfdbbbc1" containerID="620bf9c8d80335be61d4377df7f11539c5e4b98bebed586664a5842e861c97f8" exitCode=0 Sep 30 14:54:52 crc kubenswrapper[4840]: I0930 14:54:52.014316 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h4pnt" event={"ID":"4de366b7-3778-44c5-8187-4d0bdfdbbbc1","Type":"ContainerDied","Data":"620bf9c8d80335be61d4377df7f11539c5e4b98bebed586664a5842e861c97f8"} Sep 30 14:54:52 crc kubenswrapper[4840]: I0930 14:54:52.014355 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h4pnt" event={"ID":"4de366b7-3778-44c5-8187-4d0bdfdbbbc1","Type":"ContainerDied","Data":"cd337b980a30612958f844616e9119a00a5b670958c8f365ea95fbddd74e7f6f"} Sep 30 14:54:52 crc kubenswrapper[4840]: I0930 14:54:52.014388 4840 scope.go:117] "RemoveContainer" containerID="620bf9c8d80335be61d4377df7f11539c5e4b98bebed586664a5842e861c97f8" Sep 30 14:54:52 crc kubenswrapper[4840]: I0930 14:54:52.014608 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h4pnt" Sep 30 14:54:52 crc kubenswrapper[4840]: I0930 14:54:52.044066 4840 scope.go:117] "RemoveContainer" containerID="a2dfc38f8fe8ac16801edd633580c3de4ff4f7266676c5c5b8f552aec7a10e92" Sep 30 14:54:52 crc kubenswrapper[4840]: I0930 14:54:52.070348 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-h4pnt"] Sep 30 14:54:52 crc kubenswrapper[4840]: I0930 14:54:52.076269 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-h4pnt"] Sep 30 14:54:52 crc kubenswrapper[4840]: I0930 14:54:52.081467 4840 scope.go:117] "RemoveContainer" containerID="b2798b768be368e63b9ab7c81588d32e4f99494753ea7a1cbc0ee53cb1ea3ab1" Sep 30 14:54:52 crc kubenswrapper[4840]: I0930 14:54:52.118497 4840 scope.go:117] "RemoveContainer" containerID="620bf9c8d80335be61d4377df7f11539c5e4b98bebed586664a5842e861c97f8" Sep 30 14:54:52 crc kubenswrapper[4840]: E0930 14:54:52.118905 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"620bf9c8d80335be61d4377df7f11539c5e4b98bebed586664a5842e861c97f8\": container with ID starting with 620bf9c8d80335be61d4377df7f11539c5e4b98bebed586664a5842e861c97f8 not found: ID does not exist" containerID="620bf9c8d80335be61d4377df7f11539c5e4b98bebed586664a5842e861c97f8" Sep 30 14:54:52 crc kubenswrapper[4840]: I0930 14:54:52.118941 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"620bf9c8d80335be61d4377df7f11539c5e4b98bebed586664a5842e861c97f8"} err="failed to get container status \"620bf9c8d80335be61d4377df7f11539c5e4b98bebed586664a5842e861c97f8\": rpc error: code = NotFound desc = could not find container \"620bf9c8d80335be61d4377df7f11539c5e4b98bebed586664a5842e861c97f8\": container with ID starting with 620bf9c8d80335be61d4377df7f11539c5e4b98bebed586664a5842e861c97f8 not found: ID does not exist" Sep 30 14:54:52 crc kubenswrapper[4840]: I0930 14:54:52.118962 4840 scope.go:117] "RemoveContainer" containerID="a2dfc38f8fe8ac16801edd633580c3de4ff4f7266676c5c5b8f552aec7a10e92" Sep 30 14:54:52 crc kubenswrapper[4840]: E0930 14:54:52.119264 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2dfc38f8fe8ac16801edd633580c3de4ff4f7266676c5c5b8f552aec7a10e92\": container with ID starting with a2dfc38f8fe8ac16801edd633580c3de4ff4f7266676c5c5b8f552aec7a10e92 not found: ID does not exist" containerID="a2dfc38f8fe8ac16801edd633580c3de4ff4f7266676c5c5b8f552aec7a10e92" Sep 30 14:54:52 crc kubenswrapper[4840]: I0930 14:54:52.119294 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2dfc38f8fe8ac16801edd633580c3de4ff4f7266676c5c5b8f552aec7a10e92"} err="failed to get container status \"a2dfc38f8fe8ac16801edd633580c3de4ff4f7266676c5c5b8f552aec7a10e92\": rpc error: code = NotFound desc = could not find container \"a2dfc38f8fe8ac16801edd633580c3de4ff4f7266676c5c5b8f552aec7a10e92\": container with ID starting with a2dfc38f8fe8ac16801edd633580c3de4ff4f7266676c5c5b8f552aec7a10e92 not found: ID does not exist" Sep 30 14:54:52 crc kubenswrapper[4840]: I0930 14:54:52.119311 4840 scope.go:117] "RemoveContainer" containerID="b2798b768be368e63b9ab7c81588d32e4f99494753ea7a1cbc0ee53cb1ea3ab1" Sep 30 14:54:52 crc kubenswrapper[4840]: E0930 14:54:52.119568 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2798b768be368e63b9ab7c81588d32e4f99494753ea7a1cbc0ee53cb1ea3ab1\": container with ID starting with b2798b768be368e63b9ab7c81588d32e4f99494753ea7a1cbc0ee53cb1ea3ab1 not found: ID does not exist" containerID="b2798b768be368e63b9ab7c81588d32e4f99494753ea7a1cbc0ee53cb1ea3ab1" Sep 30 14:54:52 crc kubenswrapper[4840]: I0930 14:54:52.119594 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2798b768be368e63b9ab7c81588d32e4f99494753ea7a1cbc0ee53cb1ea3ab1"} err="failed to get container status \"b2798b768be368e63b9ab7c81588d32e4f99494753ea7a1cbc0ee53cb1ea3ab1\": rpc error: code = NotFound desc = could not find container \"b2798b768be368e63b9ab7c81588d32e4f99494753ea7a1cbc0ee53cb1ea3ab1\": container with ID starting with b2798b768be368e63b9ab7c81588d32e4f99494753ea7a1cbc0ee53cb1ea3ab1 not found: ID does not exist" Sep 30 14:54:52 crc kubenswrapper[4840]: I0930 14:54:52.127265 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4de366b7-3778-44c5-8187-4d0bdfdbbbc1" path="/var/lib/kubelet/pods/4de366b7-3778-44c5-8187-4d0bdfdbbbc1/volumes" Sep 30 14:55:01 crc kubenswrapper[4840]: I0930 14:55:01.098430 4840 generic.go:334] "Generic (PLEG): container finished" podID="bf9ca252-8b08-45b5-bd98-442bece17347" containerID="080b4a392d725b13e94a18caad6a5fb0fd3d342d758f311ad9ec204fa2b1a720" exitCode=0 Sep 30 14:55:01 crc kubenswrapper[4840]: I0930 14:55:01.098514 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-r4p2h/crc-debug-7zq9l" event={"ID":"bf9ca252-8b08-45b5-bd98-442bece17347","Type":"ContainerDied","Data":"080b4a392d725b13e94a18caad6a5fb0fd3d342d758f311ad9ec204fa2b1a720"} Sep 30 14:55:02 crc kubenswrapper[4840]: I0930 14:55:02.200531 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-r4p2h/crc-debug-7zq9l" Sep 30 14:55:02 crc kubenswrapper[4840]: I0930 14:55:02.229355 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-r4p2h/crc-debug-7zq9l"] Sep 30 14:55:02 crc kubenswrapper[4840]: I0930 14:55:02.235857 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-r4p2h/crc-debug-7zq9l"] Sep 30 14:55:02 crc kubenswrapper[4840]: I0930 14:55:02.389744 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbvvm\" (UniqueName: \"kubernetes.io/projected/bf9ca252-8b08-45b5-bd98-442bece17347-kube-api-access-gbvvm\") pod \"bf9ca252-8b08-45b5-bd98-442bece17347\" (UID: \"bf9ca252-8b08-45b5-bd98-442bece17347\") " Sep 30 14:55:02 crc kubenswrapper[4840]: I0930 14:55:02.389869 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bf9ca252-8b08-45b5-bd98-442bece17347-host\") pod \"bf9ca252-8b08-45b5-bd98-442bece17347\" (UID: \"bf9ca252-8b08-45b5-bd98-442bece17347\") " Sep 30 14:55:02 crc kubenswrapper[4840]: I0930 14:55:02.390326 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bf9ca252-8b08-45b5-bd98-442bece17347-host" (OuterVolumeSpecName: "host") pod "bf9ca252-8b08-45b5-bd98-442bece17347" (UID: "bf9ca252-8b08-45b5-bd98-442bece17347"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:55:02 crc kubenswrapper[4840]: I0930 14:55:02.395295 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf9ca252-8b08-45b5-bd98-442bece17347-kube-api-access-gbvvm" (OuterVolumeSpecName: "kube-api-access-gbvvm") pod "bf9ca252-8b08-45b5-bd98-442bece17347" (UID: "bf9ca252-8b08-45b5-bd98-442bece17347"). InnerVolumeSpecName "kube-api-access-gbvvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:55:02 crc kubenswrapper[4840]: I0930 14:55:02.492498 4840 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bf9ca252-8b08-45b5-bd98-442bece17347-host\") on node \"crc\" DevicePath \"\"" Sep 30 14:55:02 crc kubenswrapper[4840]: I0930 14:55:02.492748 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbvvm\" (UniqueName: \"kubernetes.io/projected/bf9ca252-8b08-45b5-bd98-442bece17347-kube-api-access-gbvvm\") on node \"crc\" DevicePath \"\"" Sep 30 14:55:03 crc kubenswrapper[4840]: I0930 14:55:03.118088 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="efd9ee9be2374911ba81c805899938fccbfe4c1e4ae0309324795953a2444d19" Sep 30 14:55:03 crc kubenswrapper[4840]: I0930 14:55:03.118121 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-r4p2h/crc-debug-7zq9l" Sep 30 14:55:03 crc kubenswrapper[4840]: I0930 14:55:03.391195 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-r4p2h/crc-debug-w4dht"] Sep 30 14:55:03 crc kubenswrapper[4840]: E0930 14:55:03.391629 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4de366b7-3778-44c5-8187-4d0bdfdbbbc1" containerName="extract-content" Sep 30 14:55:03 crc kubenswrapper[4840]: I0930 14:55:03.391646 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="4de366b7-3778-44c5-8187-4d0bdfdbbbc1" containerName="extract-content" Sep 30 14:55:03 crc kubenswrapper[4840]: E0930 14:55:03.391665 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4de366b7-3778-44c5-8187-4d0bdfdbbbc1" containerName="registry-server" Sep 30 14:55:03 crc kubenswrapper[4840]: I0930 14:55:03.391672 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="4de366b7-3778-44c5-8187-4d0bdfdbbbc1" containerName="registry-server" Sep 30 14:55:03 crc kubenswrapper[4840]: E0930 14:55:03.391726 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf9ca252-8b08-45b5-bd98-442bece17347" containerName="container-00" Sep 30 14:55:03 crc kubenswrapper[4840]: I0930 14:55:03.391736 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf9ca252-8b08-45b5-bd98-442bece17347" containerName="container-00" Sep 30 14:55:03 crc kubenswrapper[4840]: E0930 14:55:03.391760 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4de366b7-3778-44c5-8187-4d0bdfdbbbc1" containerName="extract-utilities" Sep 30 14:55:03 crc kubenswrapper[4840]: I0930 14:55:03.391768 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="4de366b7-3778-44c5-8187-4d0bdfdbbbc1" containerName="extract-utilities" Sep 30 14:55:03 crc kubenswrapper[4840]: I0930 14:55:03.391980 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf9ca252-8b08-45b5-bd98-442bece17347" containerName="container-00" Sep 30 14:55:03 crc kubenswrapper[4840]: I0930 14:55:03.391995 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="4de366b7-3778-44c5-8187-4d0bdfdbbbc1" containerName="registry-server" Sep 30 14:55:03 crc kubenswrapper[4840]: I0930 14:55:03.392700 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-r4p2h/crc-debug-w4dht" Sep 30 14:55:03 crc kubenswrapper[4840]: I0930 14:55:03.394431 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-r4p2h"/"default-dockercfg-pv7db" Sep 30 14:55:03 crc kubenswrapper[4840]: I0930 14:55:03.509511 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6183d810-3ea9-4f0c-89d1-3aaca1f4918e-host\") pod \"crc-debug-w4dht\" (UID: \"6183d810-3ea9-4f0c-89d1-3aaca1f4918e\") " pod="openshift-must-gather-r4p2h/crc-debug-w4dht" Sep 30 14:55:03 crc kubenswrapper[4840]: I0930 14:55:03.509897 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzvd7\" (UniqueName: \"kubernetes.io/projected/6183d810-3ea9-4f0c-89d1-3aaca1f4918e-kube-api-access-xzvd7\") pod \"crc-debug-w4dht\" (UID: \"6183d810-3ea9-4f0c-89d1-3aaca1f4918e\") " pod="openshift-must-gather-r4p2h/crc-debug-w4dht" Sep 30 14:55:03 crc kubenswrapper[4840]: I0930 14:55:03.611987 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzvd7\" (UniqueName: \"kubernetes.io/projected/6183d810-3ea9-4f0c-89d1-3aaca1f4918e-kube-api-access-xzvd7\") pod \"crc-debug-w4dht\" (UID: \"6183d810-3ea9-4f0c-89d1-3aaca1f4918e\") " pod="openshift-must-gather-r4p2h/crc-debug-w4dht" Sep 30 14:55:03 crc kubenswrapper[4840]: I0930 14:55:03.612093 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6183d810-3ea9-4f0c-89d1-3aaca1f4918e-host\") pod \"crc-debug-w4dht\" (UID: \"6183d810-3ea9-4f0c-89d1-3aaca1f4918e\") " pod="openshift-must-gather-r4p2h/crc-debug-w4dht" Sep 30 14:55:03 crc kubenswrapper[4840]: I0930 14:55:03.612270 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6183d810-3ea9-4f0c-89d1-3aaca1f4918e-host\") pod \"crc-debug-w4dht\" (UID: \"6183d810-3ea9-4f0c-89d1-3aaca1f4918e\") " pod="openshift-must-gather-r4p2h/crc-debug-w4dht" Sep 30 14:55:03 crc kubenswrapper[4840]: I0930 14:55:03.629385 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzvd7\" (UniqueName: \"kubernetes.io/projected/6183d810-3ea9-4f0c-89d1-3aaca1f4918e-kube-api-access-xzvd7\") pod \"crc-debug-w4dht\" (UID: \"6183d810-3ea9-4f0c-89d1-3aaca1f4918e\") " pod="openshift-must-gather-r4p2h/crc-debug-w4dht" Sep 30 14:55:03 crc kubenswrapper[4840]: I0930 14:55:03.712842 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-r4p2h/crc-debug-w4dht" Sep 30 14:55:04 crc kubenswrapper[4840]: I0930 14:55:04.128735 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf9ca252-8b08-45b5-bd98-442bece17347" path="/var/lib/kubelet/pods/bf9ca252-8b08-45b5-bd98-442bece17347/volumes" Sep 30 14:55:04 crc kubenswrapper[4840]: I0930 14:55:04.130127 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-r4p2h/crc-debug-w4dht" event={"ID":"6183d810-3ea9-4f0c-89d1-3aaca1f4918e","Type":"ContainerStarted","Data":"aa4a55fc94ddbc1c23a4fb96abd1fbde1aa569ecad5702aabc268334a51cdc8b"} Sep 30 14:55:04 crc kubenswrapper[4840]: I0930 14:55:04.130189 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-r4p2h/crc-debug-w4dht" event={"ID":"6183d810-3ea9-4f0c-89d1-3aaca1f4918e","Type":"ContainerStarted","Data":"7ef63f5f0a896ea3d7cd0f322aecd4c615a33c2871ea50aef9f39f09187bbafd"} Sep 30 14:55:04 crc kubenswrapper[4840]: I0930 14:55:04.154108 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-r4p2h/crc-debug-w4dht" podStartSLOduration=1.154087576 podStartE2EDuration="1.154087576s" podCreationTimestamp="2025-09-30 14:55:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:55:04.144758385 +0000 UTC m=+3532.773844818" watchObservedRunningTime="2025-09-30 14:55:04.154087576 +0000 UTC m=+3532.783174009" Sep 30 14:55:05 crc kubenswrapper[4840]: I0930 14:55:05.154151 4840 generic.go:334] "Generic (PLEG): container finished" podID="6183d810-3ea9-4f0c-89d1-3aaca1f4918e" containerID="aa4a55fc94ddbc1c23a4fb96abd1fbde1aa569ecad5702aabc268334a51cdc8b" exitCode=0 Sep 30 14:55:05 crc kubenswrapper[4840]: I0930 14:55:05.154197 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-r4p2h/crc-debug-w4dht" event={"ID":"6183d810-3ea9-4f0c-89d1-3aaca1f4918e","Type":"ContainerDied","Data":"aa4a55fc94ddbc1c23a4fb96abd1fbde1aa569ecad5702aabc268334a51cdc8b"} Sep 30 14:55:06 crc kubenswrapper[4840]: I0930 14:55:06.275345 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-r4p2h/crc-debug-w4dht" Sep 30 14:55:06 crc kubenswrapper[4840]: I0930 14:55:06.453214 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6183d810-3ea9-4f0c-89d1-3aaca1f4918e-host\") pod \"6183d810-3ea9-4f0c-89d1-3aaca1f4918e\" (UID: \"6183d810-3ea9-4f0c-89d1-3aaca1f4918e\") " Sep 30 14:55:06 crc kubenswrapper[4840]: I0930 14:55:06.453373 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzvd7\" (UniqueName: \"kubernetes.io/projected/6183d810-3ea9-4f0c-89d1-3aaca1f4918e-kube-api-access-xzvd7\") pod \"6183d810-3ea9-4f0c-89d1-3aaca1f4918e\" (UID: \"6183d810-3ea9-4f0c-89d1-3aaca1f4918e\") " Sep 30 14:55:06 crc kubenswrapper[4840]: I0930 14:55:06.453812 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6183d810-3ea9-4f0c-89d1-3aaca1f4918e-host" (OuterVolumeSpecName: "host") pod "6183d810-3ea9-4f0c-89d1-3aaca1f4918e" (UID: "6183d810-3ea9-4f0c-89d1-3aaca1f4918e"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:55:06 crc kubenswrapper[4840]: I0930 14:55:06.466592 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6183d810-3ea9-4f0c-89d1-3aaca1f4918e-kube-api-access-xzvd7" (OuterVolumeSpecName: "kube-api-access-xzvd7") pod "6183d810-3ea9-4f0c-89d1-3aaca1f4918e" (UID: "6183d810-3ea9-4f0c-89d1-3aaca1f4918e"). InnerVolumeSpecName "kube-api-access-xzvd7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:55:06 crc kubenswrapper[4840]: I0930 14:55:06.555882 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzvd7\" (UniqueName: \"kubernetes.io/projected/6183d810-3ea9-4f0c-89d1-3aaca1f4918e-kube-api-access-xzvd7\") on node \"crc\" DevicePath \"\"" Sep 30 14:55:06 crc kubenswrapper[4840]: I0930 14:55:06.555914 4840 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6183d810-3ea9-4f0c-89d1-3aaca1f4918e-host\") on node \"crc\" DevicePath \"\"" Sep 30 14:55:07 crc kubenswrapper[4840]: I0930 14:55:07.180288 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-r4p2h/crc-debug-w4dht" event={"ID":"6183d810-3ea9-4f0c-89d1-3aaca1f4918e","Type":"ContainerDied","Data":"7ef63f5f0a896ea3d7cd0f322aecd4c615a33c2871ea50aef9f39f09187bbafd"} Sep 30 14:55:07 crc kubenswrapper[4840]: I0930 14:55:07.180649 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ef63f5f0a896ea3d7cd0f322aecd4c615a33c2871ea50aef9f39f09187bbafd" Sep 30 14:55:07 crc kubenswrapper[4840]: I0930 14:55:07.180338 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-r4p2h/crc-debug-w4dht" Sep 30 14:55:10 crc kubenswrapper[4840]: I0930 14:55:10.556509 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-r4p2h/crc-debug-w4dht"] Sep 30 14:55:10 crc kubenswrapper[4840]: I0930 14:55:10.565353 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-r4p2h/crc-debug-w4dht"] Sep 30 14:55:11 crc kubenswrapper[4840]: I0930 14:55:11.710638 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-r4p2h/crc-debug-bzxv8"] Sep 30 14:55:11 crc kubenswrapper[4840]: E0930 14:55:11.710983 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6183d810-3ea9-4f0c-89d1-3aaca1f4918e" containerName="container-00" Sep 30 14:55:11 crc kubenswrapper[4840]: I0930 14:55:11.710996 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="6183d810-3ea9-4f0c-89d1-3aaca1f4918e" containerName="container-00" Sep 30 14:55:11 crc kubenswrapper[4840]: I0930 14:55:11.711204 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="6183d810-3ea9-4f0c-89d1-3aaca1f4918e" containerName="container-00" Sep 30 14:55:11 crc kubenswrapper[4840]: I0930 14:55:11.711785 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-r4p2h/crc-debug-bzxv8" Sep 30 14:55:11 crc kubenswrapper[4840]: I0930 14:55:11.713833 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-r4p2h"/"default-dockercfg-pv7db" Sep 30 14:55:11 crc kubenswrapper[4840]: I0930 14:55:11.840752 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dn226\" (UniqueName: \"kubernetes.io/projected/510e3570-29ff-4148-8314-a4707f1af37a-kube-api-access-dn226\") pod \"crc-debug-bzxv8\" (UID: \"510e3570-29ff-4148-8314-a4707f1af37a\") " pod="openshift-must-gather-r4p2h/crc-debug-bzxv8" Sep 30 14:55:11 crc kubenswrapper[4840]: I0930 14:55:11.840832 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/510e3570-29ff-4148-8314-a4707f1af37a-host\") pod \"crc-debug-bzxv8\" (UID: \"510e3570-29ff-4148-8314-a4707f1af37a\") " pod="openshift-must-gather-r4p2h/crc-debug-bzxv8" Sep 30 14:55:11 crc kubenswrapper[4840]: I0930 14:55:11.942689 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dn226\" (UniqueName: \"kubernetes.io/projected/510e3570-29ff-4148-8314-a4707f1af37a-kube-api-access-dn226\") pod \"crc-debug-bzxv8\" (UID: \"510e3570-29ff-4148-8314-a4707f1af37a\") " pod="openshift-must-gather-r4p2h/crc-debug-bzxv8" Sep 30 14:55:11 crc kubenswrapper[4840]: I0930 14:55:11.943058 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/510e3570-29ff-4148-8314-a4707f1af37a-host\") pod \"crc-debug-bzxv8\" (UID: \"510e3570-29ff-4148-8314-a4707f1af37a\") " pod="openshift-must-gather-r4p2h/crc-debug-bzxv8" Sep 30 14:55:11 crc kubenswrapper[4840]: I0930 14:55:11.943187 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/510e3570-29ff-4148-8314-a4707f1af37a-host\") pod \"crc-debug-bzxv8\" (UID: \"510e3570-29ff-4148-8314-a4707f1af37a\") " pod="openshift-must-gather-r4p2h/crc-debug-bzxv8" Sep 30 14:55:11 crc kubenswrapper[4840]: I0930 14:55:11.967138 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dn226\" (UniqueName: \"kubernetes.io/projected/510e3570-29ff-4148-8314-a4707f1af37a-kube-api-access-dn226\") pod \"crc-debug-bzxv8\" (UID: \"510e3570-29ff-4148-8314-a4707f1af37a\") " pod="openshift-must-gather-r4p2h/crc-debug-bzxv8" Sep 30 14:55:12 crc kubenswrapper[4840]: I0930 14:55:12.037401 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-r4p2h"/"default-dockercfg-pv7db" Sep 30 14:55:12 crc kubenswrapper[4840]: I0930 14:55:12.045612 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-r4p2h/crc-debug-bzxv8" Sep 30 14:55:12 crc kubenswrapper[4840]: I0930 14:55:12.130801 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6183d810-3ea9-4f0c-89d1-3aaca1f4918e" path="/var/lib/kubelet/pods/6183d810-3ea9-4f0c-89d1-3aaca1f4918e/volumes" Sep 30 14:55:12 crc kubenswrapper[4840]: I0930 14:55:12.226224 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-r4p2h/crc-debug-bzxv8" event={"ID":"510e3570-29ff-4148-8314-a4707f1af37a","Type":"ContainerStarted","Data":"ac00c81845abe4273fd7e1e48e8bceeb8d7c4698c4e56537bc930267c44ef31d"} Sep 30 14:55:13 crc kubenswrapper[4840]: I0930 14:55:13.238364 4840 generic.go:334] "Generic (PLEG): container finished" podID="510e3570-29ff-4148-8314-a4707f1af37a" containerID="af771f4ba2b0430bf10d7e20f490856ef6525b05dacc1eb7bedba7d959a3cd3c" exitCode=0 Sep 30 14:55:13 crc kubenswrapper[4840]: I0930 14:55:13.238482 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-r4p2h/crc-debug-bzxv8" event={"ID":"510e3570-29ff-4148-8314-a4707f1af37a","Type":"ContainerDied","Data":"af771f4ba2b0430bf10d7e20f490856ef6525b05dacc1eb7bedba7d959a3cd3c"} Sep 30 14:55:13 crc kubenswrapper[4840]: I0930 14:55:13.274651 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-r4p2h/crc-debug-bzxv8"] Sep 30 14:55:13 crc kubenswrapper[4840]: I0930 14:55:13.285055 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-r4p2h/crc-debug-bzxv8"] Sep 30 14:55:14 crc kubenswrapper[4840]: I0930 14:55:14.358537 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-r4p2h/crc-debug-bzxv8" Sep 30 14:55:14 crc kubenswrapper[4840]: I0930 14:55:14.489104 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dn226\" (UniqueName: \"kubernetes.io/projected/510e3570-29ff-4148-8314-a4707f1af37a-kube-api-access-dn226\") pod \"510e3570-29ff-4148-8314-a4707f1af37a\" (UID: \"510e3570-29ff-4148-8314-a4707f1af37a\") " Sep 30 14:55:14 crc kubenswrapper[4840]: I0930 14:55:14.489152 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/510e3570-29ff-4148-8314-a4707f1af37a-host\") pod \"510e3570-29ff-4148-8314-a4707f1af37a\" (UID: \"510e3570-29ff-4148-8314-a4707f1af37a\") " Sep 30 14:55:14 crc kubenswrapper[4840]: I0930 14:55:14.489305 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/510e3570-29ff-4148-8314-a4707f1af37a-host" (OuterVolumeSpecName: "host") pod "510e3570-29ff-4148-8314-a4707f1af37a" (UID: "510e3570-29ff-4148-8314-a4707f1af37a"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:55:14 crc kubenswrapper[4840]: I0930 14:55:14.489669 4840 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/510e3570-29ff-4148-8314-a4707f1af37a-host\") on node \"crc\" DevicePath \"\"" Sep 30 14:55:14 crc kubenswrapper[4840]: I0930 14:55:14.495436 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/510e3570-29ff-4148-8314-a4707f1af37a-kube-api-access-dn226" (OuterVolumeSpecName: "kube-api-access-dn226") pod "510e3570-29ff-4148-8314-a4707f1af37a" (UID: "510e3570-29ff-4148-8314-a4707f1af37a"). InnerVolumeSpecName "kube-api-access-dn226". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:55:14 crc kubenswrapper[4840]: I0930 14:55:14.590814 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dn226\" (UniqueName: \"kubernetes.io/projected/510e3570-29ff-4148-8314-a4707f1af37a-kube-api-access-dn226\") on node \"crc\" DevicePath \"\"" Sep 30 14:55:14 crc kubenswrapper[4840]: I0930 14:55:14.688976 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx_181ca13b-195f-45ea-9894-3383d192d377/util/0.log" Sep 30 14:55:14 crc kubenswrapper[4840]: I0930 14:55:14.836604 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx_181ca13b-195f-45ea-9894-3383d192d377/util/0.log" Sep 30 14:55:14 crc kubenswrapper[4840]: I0930 14:55:14.869590 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx_181ca13b-195f-45ea-9894-3383d192d377/pull/0.log" Sep 30 14:55:14 crc kubenswrapper[4840]: I0930 14:55:14.869691 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx_181ca13b-195f-45ea-9894-3383d192d377/pull/0.log" Sep 30 14:55:15 crc kubenswrapper[4840]: I0930 14:55:15.006677 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx_181ca13b-195f-45ea-9894-3383d192d377/util/0.log" Sep 30 14:55:15 crc kubenswrapper[4840]: I0930 14:55:15.027927 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx_181ca13b-195f-45ea-9894-3383d192d377/pull/0.log" Sep 30 14:55:15 crc kubenswrapper[4840]: I0930 14:55:15.036907 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx_181ca13b-195f-45ea-9894-3383d192d377/extract/0.log" Sep 30 14:55:15 crc kubenswrapper[4840]: I0930 14:55:15.178491 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-rnsbn_0f4e505b-4892-4c92-9ec4-0bdeb5d4c873/kube-rbac-proxy/0.log" Sep 30 14:55:15 crc kubenswrapper[4840]: I0930 14:55:15.234503 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-rnsbn_0f4e505b-4892-4c92-9ec4-0bdeb5d4c873/manager/0.log" Sep 30 14:55:15 crc kubenswrapper[4840]: I0930 14:55:15.266176 4840 scope.go:117] "RemoveContainer" containerID="af771f4ba2b0430bf10d7e20f490856ef6525b05dacc1eb7bedba7d959a3cd3c" Sep 30 14:55:15 crc kubenswrapper[4840]: I0930 14:55:15.266389 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-r4p2h/crc-debug-bzxv8" Sep 30 14:55:15 crc kubenswrapper[4840]: I0930 14:55:15.280435 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-smg6p_3ffb0dd0-3dd8-4f37-8ead-4add746a1fc6/kube-rbac-proxy/0.log" Sep 30 14:55:15 crc kubenswrapper[4840]: I0930 14:55:15.419566 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-6nn29_395025e2-9d12-4172-bb73-b4fb67e0111c/kube-rbac-proxy/0.log" Sep 30 14:55:15 crc kubenswrapper[4840]: I0930 14:55:15.434325 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-smg6p_3ffb0dd0-3dd8-4f37-8ead-4add746a1fc6/manager/0.log" Sep 30 14:55:15 crc kubenswrapper[4840]: I0930 14:55:15.466814 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-6nn29_395025e2-9d12-4172-bb73-b4fb67e0111c/manager/0.log" Sep 30 14:55:15 crc kubenswrapper[4840]: I0930 14:55:15.592165 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-sq5hd_64c5ad91-f113-46e3-ac87-36521d32f4f0/kube-rbac-proxy/0.log" Sep 30 14:55:15 crc kubenswrapper[4840]: I0930 14:55:15.715993 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-sq5hd_64c5ad91-f113-46e3-ac87-36521d32f4f0/manager/0.log" Sep 30 14:55:15 crc kubenswrapper[4840]: I0930 14:55:15.774545 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-jwlj9_5daf5662-8848-4fd9-a6af-e091fe303d89/manager/0.log" Sep 30 14:55:15 crc kubenswrapper[4840]: I0930 14:55:15.799789 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-jwlj9_5daf5662-8848-4fd9-a6af-e091fe303d89/kube-rbac-proxy/0.log" Sep 30 14:55:15 crc kubenswrapper[4840]: I0930 14:55:15.906800 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-d9jgv_63d94c36-0a55-472c-bbe8-27195105584e/kube-rbac-proxy/0.log" Sep 30 14:55:15 crc kubenswrapper[4840]: I0930 14:55:15.996735 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-d9jgv_63d94c36-0a55-472c-bbe8-27195105584e/manager/0.log" Sep 30 14:55:16 crc kubenswrapper[4840]: I0930 14:55:16.041119 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-7d857cc749-xtqwd_807f2091-02fa-41e0-943a-b30f4202c476/kube-rbac-proxy/0.log" Sep 30 14:55:16 crc kubenswrapper[4840]: I0930 14:55:16.128731 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="510e3570-29ff-4148-8314-a4707f1af37a" path="/var/lib/kubelet/pods/510e3570-29ff-4148-8314-a4707f1af37a/volumes" Sep 30 14:55:16 crc kubenswrapper[4840]: I0930 14:55:16.185605 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-7975b88857-fg7c7_494a382c-6fec-4711-8a90-7f691c847e86/kube-rbac-proxy/0.log" Sep 30 14:55:16 crc kubenswrapper[4840]: I0930 14:55:16.222258 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-7975b88857-fg7c7_494a382c-6fec-4711-8a90-7f691c847e86/manager/0.log" Sep 30 14:55:16 crc kubenswrapper[4840]: I0930 14:55:16.229784 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-7d857cc749-xtqwd_807f2091-02fa-41e0-943a-b30f4202c476/manager/0.log" Sep 30 14:55:16 crc kubenswrapper[4840]: I0930 14:55:16.361602 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5bd55b4bff-rjpbm_b927784b-0b44-4bf1-9751-3aafcc24367b/kube-rbac-proxy/0.log" Sep 30 14:55:16 crc kubenswrapper[4840]: I0930 14:55:16.432438 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5bd55b4bff-rjpbm_b927784b-0b44-4bf1-9751-3aafcc24367b/manager/0.log" Sep 30 14:55:16 crc kubenswrapper[4840]: I0930 14:55:16.546945 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-d2hgd_eb939f27-ce94-408d-b673-4c25f96dd059/kube-rbac-proxy/0.log" Sep 30 14:55:16 crc kubenswrapper[4840]: I0930 14:55:16.557009 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-d2hgd_eb939f27-ce94-408d-b673-4c25f96dd059/manager/0.log" Sep 30 14:55:16 crc kubenswrapper[4840]: I0930 14:55:16.644013 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-xzmg8_385365f0-9e2c-4c89-b95b-164a473ff0c3/kube-rbac-proxy/0.log" Sep 30 14:55:16 crc kubenswrapper[4840]: I0930 14:55:16.744093 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-xzmg8_385365f0-9e2c-4c89-b95b-164a473ff0c3/manager/0.log" Sep 30 14:55:16 crc kubenswrapper[4840]: I0930 14:55:16.806410 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-64d7b59854-qsq84_3d5da746-e5f3-4bed-9c77-37a78da8f1a6/kube-rbac-proxy/0.log" Sep 30 14:55:16 crc kubenswrapper[4840]: I0930 14:55:16.876978 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-64d7b59854-qsq84_3d5da746-e5f3-4bed-9c77-37a78da8f1a6/manager/0.log" Sep 30 14:55:16 crc kubenswrapper[4840]: I0930 14:55:16.937926 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-c7c776c96-x4mx4_3f4d112b-e6b8-455a-9f66-e8ee1bb81ac9/kube-rbac-proxy/0.log" Sep 30 14:55:17 crc kubenswrapper[4840]: I0930 14:55:17.046039 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-c7c776c96-x4mx4_3f4d112b-e6b8-455a-9f66-e8ee1bb81ac9/manager/0.log" Sep 30 14:55:17 crc kubenswrapper[4840]: I0930 14:55:17.148538 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-76fcc6dc7c-t2mx4_a76ab4ec-b180-473f-a597-641725526b58/manager/0.log" Sep 30 14:55:17 crc kubenswrapper[4840]: I0930 14:55:17.150228 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-76fcc6dc7c-t2mx4_a76ab4ec-b180-473f-a597-641725526b58/kube-rbac-proxy/0.log" Sep 30 14:55:17 crc kubenswrapper[4840]: I0930 14:55:17.260882 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6d776955-p6msl_62475ec3-b883-4d16-bb31-6ee393062da2/kube-rbac-proxy/0.log" Sep 30 14:55:17 crc kubenswrapper[4840]: I0930 14:55:17.327348 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6d776955-p6msl_62475ec3-b883-4d16-bb31-6ee393062da2/manager/0.log" Sep 30 14:55:17 crc kubenswrapper[4840]: I0930 14:55:17.449077 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-568796b78d-bvbdn_e61dcaab-7705-4857-bdd9-adeb07d9e6f2/kube-rbac-proxy/0.log" Sep 30 14:55:17 crc kubenswrapper[4840]: I0930 14:55:17.615398 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-84594b7f86-xwqzh_0b30b057-e591-411e-919b-5d1efac1a249/kube-rbac-proxy/0.log" Sep 30 14:55:17 crc kubenswrapper[4840]: I0930 14:55:17.780651 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-84594b7f86-xwqzh_0b30b057-e591-411e-919b-5d1efac1a249/operator/0.log" Sep 30 14:55:17 crc kubenswrapper[4840]: I0930 14:55:17.826274 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-gl8hl_a1690286-e0ea-4cf7-877b-d5e172dc365e/registry-server/0.log" Sep 30 14:55:18 crc kubenswrapper[4840]: I0930 14:55:18.004789 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-9t25m_dc75c0e8-691e-4108-b9c7-e4233d0baeb6/kube-rbac-proxy/0.log" Sep 30 14:55:18 crc kubenswrapper[4840]: I0930 14:55:18.099387 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-9t25m_dc75c0e8-691e-4108-b9c7-e4233d0baeb6/manager/0.log" Sep 30 14:55:18 crc kubenswrapper[4840]: I0930 14:55:18.347186 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-tqjvl_3ed332eb-b604-4667-90f4-541d5a9115c2/manager/0.log" Sep 30 14:55:18 crc kubenswrapper[4840]: I0930 14:55:18.355488 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-tqjvl_3ed332eb-b604-4667-90f4-541d5a9115c2/kube-rbac-proxy/0.log" Sep 30 14:55:18 crc kubenswrapper[4840]: I0930 14:55:18.489580 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-79d8469568-smjg7_aa2cf241-ec41-4a94-b452-653b3439a073/operator/0.log" Sep 30 14:55:18 crc kubenswrapper[4840]: I0930 14:55:18.546454 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-568796b78d-bvbdn_e61dcaab-7705-4857-bdd9-adeb07d9e6f2/manager/0.log" Sep 30 14:55:18 crc kubenswrapper[4840]: I0930 14:55:18.567447 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-bc7dc7bd9-4mpb8_2b8703c9-f691-4c3a-9941-8bcd36aaf948/manager/0.log" Sep 30 14:55:18 crc kubenswrapper[4840]: I0930 14:55:18.593665 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-bc7dc7bd9-4mpb8_2b8703c9-f691-4c3a-9941-8bcd36aaf948/kube-rbac-proxy/0.log" Sep 30 14:55:18 crc kubenswrapper[4840]: I0930 14:55:18.729409 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-78974455b9-97rs4_47248483-6556-491b-b65f-01cad9f1d1fc/kube-rbac-proxy/0.log" Sep 30 14:55:18 crc kubenswrapper[4840]: I0930 14:55:18.828213 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-f66b554c6-9npm5_3ba6da78-9606-4e1c-bc55-6cbe2e052464/kube-rbac-proxy/0.log" Sep 30 14:55:18 crc kubenswrapper[4840]: I0930 14:55:18.833629 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-78974455b9-97rs4_47248483-6556-491b-b65f-01cad9f1d1fc/manager/0.log" Sep 30 14:55:18 crc kubenswrapper[4840]: I0930 14:55:18.955187 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-f66b554c6-9npm5_3ba6da78-9606-4e1c-bc55-6cbe2e052464/manager/0.log" Sep 30 14:55:19 crc kubenswrapper[4840]: I0930 14:55:19.002813 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-76669f99c-rn2qq_e1cd519e-b5c3-48f2-b426-d0f4e912683b/kube-rbac-proxy/0.log" Sep 30 14:55:19 crc kubenswrapper[4840]: I0930 14:55:19.029584 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-76669f99c-rn2qq_e1cd519e-b5c3-48f2-b426-d0f4e912683b/manager/0.log" Sep 30 14:55:32 crc kubenswrapper[4840]: I0930 14:55:32.333154 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-z866q_eb59a261-a39c-4be5-86f0-a14643aef960/control-plane-machine-set-operator/0.log" Sep 30 14:55:32 crc kubenswrapper[4840]: I0930 14:55:32.547177 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-zbnls_d3508091-c5c6-4dca-9b32-883426154a6e/kube-rbac-proxy/0.log" Sep 30 14:55:32 crc kubenswrapper[4840]: I0930 14:55:32.588375 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-zbnls_d3508091-c5c6-4dca-9b32-883426154a6e/machine-api-operator/0.log" Sep 30 14:55:42 crc kubenswrapper[4840]: I0930 14:55:42.818034 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-hjfs2_11b92f8e-b36f-47e5-8d1a-d3c83ccbd870/cert-manager-controller/0.log" Sep 30 14:55:42 crc kubenswrapper[4840]: I0930 14:55:42.994213 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-f4kpb_8df1331d-54ff-47cd-80c9-c18a5ab2df53/cert-manager-cainjector/0.log" Sep 30 14:55:43 crc kubenswrapper[4840]: I0930 14:55:43.017396 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-sszwr_18d21186-8397-400e-a9df-5e4e41e6739e/cert-manager-webhook/0.log" Sep 30 14:55:51 crc kubenswrapper[4840]: I0930 14:55:51.871837 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:55:51 crc kubenswrapper[4840]: I0930 14:55:51.872415 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:55:53 crc kubenswrapper[4840]: I0930 14:55:53.351465 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-864bb6dfb5-lw82x_14ccc5bb-b26a-42df-b0b9-f1ea9bcbff31/nmstate-console-plugin/0.log" Sep 30 14:55:53 crc kubenswrapper[4840]: I0930 14:55:53.481442 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-gnz9p_7e3e89d5-05d9-4675-9dad-83f59258f876/nmstate-handler/0.log" Sep 30 14:55:53 crc kubenswrapper[4840]: I0930 14:55:53.512941 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58fcddf996-bw56r_9ed06022-f9fe-4b61-bd42-ed308118a717/kube-rbac-proxy/0.log" Sep 30 14:55:53 crc kubenswrapper[4840]: I0930 14:55:53.540721 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58fcddf996-bw56r_9ed06022-f9fe-4b61-bd42-ed308118a717/nmstate-metrics/0.log" Sep 30 14:55:53 crc kubenswrapper[4840]: I0930 14:55:53.682302 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5d6f6cfd66-wqtsv_5665a62e-e1b7-4773-afcb-009f7ffb354d/nmstate-operator/0.log" Sep 30 14:55:53 crc kubenswrapper[4840]: I0930 14:55:53.727609 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6d689559c5-h7fmf_e047fa81-9835-4bca-95c4-c8d24f97857b/nmstate-webhook/0.log" Sep 30 14:56:06 crc kubenswrapper[4840]: I0930 14:56:06.319333 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5d688f5ffc-nv2dz_e2682a09-9f63-42a1-882a-368be12b8be0/kube-rbac-proxy/0.log" Sep 30 14:56:06 crc kubenswrapper[4840]: I0930 14:56:06.428001 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5d688f5ffc-nv2dz_e2682a09-9f63-42a1-882a-368be12b8be0/controller/0.log" Sep 30 14:56:06 crc kubenswrapper[4840]: I0930 14:56:06.522055 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sqx2j_624a00a9-8fda-4e40-9df0-e8452ed89d28/cp-frr-files/0.log" Sep 30 14:56:06 crc kubenswrapper[4840]: I0930 14:56:06.674505 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sqx2j_624a00a9-8fda-4e40-9df0-e8452ed89d28/cp-frr-files/0.log" Sep 30 14:56:06 crc kubenswrapper[4840]: I0930 14:56:06.713619 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sqx2j_624a00a9-8fda-4e40-9df0-e8452ed89d28/cp-metrics/0.log" Sep 30 14:56:06 crc kubenswrapper[4840]: I0930 14:56:06.906603 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sqx2j_624a00a9-8fda-4e40-9df0-e8452ed89d28/cp-reloader/0.log" Sep 30 14:56:06 crc kubenswrapper[4840]: I0930 14:56:06.959431 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sqx2j_624a00a9-8fda-4e40-9df0-e8452ed89d28/cp-reloader/0.log" Sep 30 14:56:07 crc kubenswrapper[4840]: I0930 14:56:07.129901 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sqx2j_624a00a9-8fda-4e40-9df0-e8452ed89d28/cp-frr-files/0.log" Sep 30 14:56:07 crc kubenswrapper[4840]: I0930 14:56:07.167749 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sqx2j_624a00a9-8fda-4e40-9df0-e8452ed89d28/cp-metrics/0.log" Sep 30 14:56:07 crc kubenswrapper[4840]: I0930 14:56:07.189180 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sqx2j_624a00a9-8fda-4e40-9df0-e8452ed89d28/cp-reloader/0.log" Sep 30 14:56:07 crc kubenswrapper[4840]: I0930 14:56:07.190324 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sqx2j_624a00a9-8fda-4e40-9df0-e8452ed89d28/cp-metrics/0.log" Sep 30 14:56:07 crc kubenswrapper[4840]: I0930 14:56:07.347486 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sqx2j_624a00a9-8fda-4e40-9df0-e8452ed89d28/cp-reloader/0.log" Sep 30 14:56:07 crc kubenswrapper[4840]: I0930 14:56:07.347645 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sqx2j_624a00a9-8fda-4e40-9df0-e8452ed89d28/cp-frr-files/0.log" Sep 30 14:56:07 crc kubenswrapper[4840]: I0930 14:56:07.386647 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sqx2j_624a00a9-8fda-4e40-9df0-e8452ed89d28/cp-metrics/0.log" Sep 30 14:56:07 crc kubenswrapper[4840]: I0930 14:56:07.407387 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sqx2j_624a00a9-8fda-4e40-9df0-e8452ed89d28/controller/0.log" Sep 30 14:56:07 crc kubenswrapper[4840]: I0930 14:56:07.530416 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sqx2j_624a00a9-8fda-4e40-9df0-e8452ed89d28/frr-metrics/0.log" Sep 30 14:56:07 crc kubenswrapper[4840]: I0930 14:56:07.615505 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sqx2j_624a00a9-8fda-4e40-9df0-e8452ed89d28/kube-rbac-proxy/0.log" Sep 30 14:56:07 crc kubenswrapper[4840]: I0930 14:56:07.647248 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sqx2j_624a00a9-8fda-4e40-9df0-e8452ed89d28/kube-rbac-proxy-frr/0.log" Sep 30 14:56:07 crc kubenswrapper[4840]: I0930 14:56:07.799656 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sqx2j_624a00a9-8fda-4e40-9df0-e8452ed89d28/reloader/0.log" Sep 30 14:56:07 crc kubenswrapper[4840]: I0930 14:56:07.900326 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-5478bdb765-8qmmm_7348a607-eb78-4fbf-a620-9be8180b8503/frr-k8s-webhook-server/0.log" Sep 30 14:56:08 crc kubenswrapper[4840]: I0930 14:56:08.083405 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6c8fd557c5-29r6x_ec2b2b7b-3b95-42c0-bfde-d8206541fe62/manager/0.log" Sep 30 14:56:08 crc kubenswrapper[4840]: I0930 14:56:08.270645 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-79c987c9f-v9dzf_1e99fa8d-4919-483b-a71e-ee137198e55c/webhook-server/0.log" Sep 30 14:56:08 crc kubenswrapper[4840]: I0930 14:56:08.370244 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-hnv4n_a15e1237-7226-4180-8c27-fb06dffbd2d7/kube-rbac-proxy/0.log" Sep 30 14:56:08 crc kubenswrapper[4840]: I0930 14:56:08.810884 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sqx2j_624a00a9-8fda-4e40-9df0-e8452ed89d28/frr/0.log" Sep 30 14:56:08 crc kubenswrapper[4840]: I0930 14:56:08.908891 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-hnv4n_a15e1237-7226-4180-8c27-fb06dffbd2d7/speaker/0.log" Sep 30 14:56:19 crc kubenswrapper[4840]: I0930 14:56:19.365278 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6_0baa1415-f354-45dd-a352-140902beba82/util/0.log" Sep 30 14:56:19 crc kubenswrapper[4840]: I0930 14:56:19.539020 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6_0baa1415-f354-45dd-a352-140902beba82/pull/0.log" Sep 30 14:56:19 crc kubenswrapper[4840]: I0930 14:56:19.543078 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6_0baa1415-f354-45dd-a352-140902beba82/util/0.log" Sep 30 14:56:19 crc kubenswrapper[4840]: I0930 14:56:19.550728 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6_0baa1415-f354-45dd-a352-140902beba82/pull/0.log" Sep 30 14:56:19 crc kubenswrapper[4840]: I0930 14:56:19.694445 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6_0baa1415-f354-45dd-a352-140902beba82/extract/0.log" Sep 30 14:56:19 crc kubenswrapper[4840]: I0930 14:56:19.717208 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6_0baa1415-f354-45dd-a352-140902beba82/pull/0.log" Sep 30 14:56:19 crc kubenswrapper[4840]: I0930 14:56:19.720048 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6_0baa1415-f354-45dd-a352-140902beba82/util/0.log" Sep 30 14:56:19 crc kubenswrapper[4840]: I0930 14:56:19.861313 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zbb8b_233f29d4-5cbe-4cd1-b8ac-e40cf332d381/extract-utilities/0.log" Sep 30 14:56:20 crc kubenswrapper[4840]: I0930 14:56:20.066202 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zbb8b_233f29d4-5cbe-4cd1-b8ac-e40cf332d381/extract-utilities/0.log" Sep 30 14:56:20 crc kubenswrapper[4840]: I0930 14:56:20.067677 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zbb8b_233f29d4-5cbe-4cd1-b8ac-e40cf332d381/extract-content/0.log" Sep 30 14:56:20 crc kubenswrapper[4840]: I0930 14:56:20.091064 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zbb8b_233f29d4-5cbe-4cd1-b8ac-e40cf332d381/extract-content/0.log" Sep 30 14:56:20 crc kubenswrapper[4840]: I0930 14:56:20.313824 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zbb8b_233f29d4-5cbe-4cd1-b8ac-e40cf332d381/extract-utilities/0.log" Sep 30 14:56:20 crc kubenswrapper[4840]: I0930 14:56:20.334831 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zbb8b_233f29d4-5cbe-4cd1-b8ac-e40cf332d381/extract-content/0.log" Sep 30 14:56:20 crc kubenswrapper[4840]: I0930 14:56:20.545112 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-t2tp4_e5a51075-4798-48bf-ac82-67a642d1269a/extract-utilities/0.log" Sep 30 14:56:20 crc kubenswrapper[4840]: I0930 14:56:20.785418 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-t2tp4_e5a51075-4798-48bf-ac82-67a642d1269a/extract-content/0.log" Sep 30 14:56:20 crc kubenswrapper[4840]: I0930 14:56:20.797013 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-t2tp4_e5a51075-4798-48bf-ac82-67a642d1269a/extract-utilities/0.log" Sep 30 14:56:20 crc kubenswrapper[4840]: I0930 14:56:20.812359 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zbb8b_233f29d4-5cbe-4cd1-b8ac-e40cf332d381/registry-server/0.log" Sep 30 14:56:20 crc kubenswrapper[4840]: I0930 14:56:20.857156 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-t2tp4_e5a51075-4798-48bf-ac82-67a642d1269a/extract-content/0.log" Sep 30 14:56:20 crc kubenswrapper[4840]: I0930 14:56:20.998534 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-t2tp4_e5a51075-4798-48bf-ac82-67a642d1269a/extract-content/0.log" Sep 30 14:56:21 crc kubenswrapper[4840]: I0930 14:56:21.020932 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-t2tp4_e5a51075-4798-48bf-ac82-67a642d1269a/extract-utilities/0.log" Sep 30 14:56:21 crc kubenswrapper[4840]: I0930 14:56:21.268720 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j_e8ac07f8-b0b5-4ff0-bae8-4d6594567c32/util/0.log" Sep 30 14:56:21 crc kubenswrapper[4840]: I0930 14:56:21.436615 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j_e8ac07f8-b0b5-4ff0-bae8-4d6594567c32/util/0.log" Sep 30 14:56:21 crc kubenswrapper[4840]: I0930 14:56:21.527455 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j_e8ac07f8-b0b5-4ff0-bae8-4d6594567c32/pull/0.log" Sep 30 14:56:21 crc kubenswrapper[4840]: I0930 14:56:21.527760 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-t2tp4_e5a51075-4798-48bf-ac82-67a642d1269a/registry-server/0.log" Sep 30 14:56:21 crc kubenswrapper[4840]: I0930 14:56:21.539175 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j_e8ac07f8-b0b5-4ff0-bae8-4d6594567c32/pull/0.log" Sep 30 14:56:21 crc kubenswrapper[4840]: I0930 14:56:21.687325 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j_e8ac07f8-b0b5-4ff0-bae8-4d6594567c32/pull/0.log" Sep 30 14:56:21 crc kubenswrapper[4840]: I0930 14:56:21.694321 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j_e8ac07f8-b0b5-4ff0-bae8-4d6594567c32/util/0.log" Sep 30 14:56:21 crc kubenswrapper[4840]: I0930 14:56:21.746862 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j_e8ac07f8-b0b5-4ff0-bae8-4d6594567c32/extract/0.log" Sep 30 14:56:21 crc kubenswrapper[4840]: I0930 14:56:21.853043 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-hfdnb_b0c00d91-ae86-48df-84b1-1e1c29b7ff38/marketplace-operator/0.log" Sep 30 14:56:21 crc kubenswrapper[4840]: I0930 14:56:21.871587 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:56:21 crc kubenswrapper[4840]: I0930 14:56:21.871680 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:56:21 crc kubenswrapper[4840]: I0930 14:56:21.938920 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f85rt_0207a121-1290-431e-85a5-e5bfc5355d50/extract-utilities/0.log" Sep 30 14:56:22 crc kubenswrapper[4840]: I0930 14:56:22.114485 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f85rt_0207a121-1290-431e-85a5-e5bfc5355d50/extract-content/0.log" Sep 30 14:56:22 crc kubenswrapper[4840]: I0930 14:56:22.124604 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f85rt_0207a121-1290-431e-85a5-e5bfc5355d50/extract-utilities/0.log" Sep 30 14:56:22 crc kubenswrapper[4840]: I0930 14:56:22.141445 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f85rt_0207a121-1290-431e-85a5-e5bfc5355d50/extract-content/0.log" Sep 30 14:56:22 crc kubenswrapper[4840]: I0930 14:56:22.301410 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f85rt_0207a121-1290-431e-85a5-e5bfc5355d50/extract-utilities/0.log" Sep 30 14:56:22 crc kubenswrapper[4840]: I0930 14:56:22.341874 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f85rt_0207a121-1290-431e-85a5-e5bfc5355d50/extract-content/0.log" Sep 30 14:56:22 crc kubenswrapper[4840]: I0930 14:56:22.455830 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f85rt_0207a121-1290-431e-85a5-e5bfc5355d50/registry-server/0.log" Sep 30 14:56:22 crc kubenswrapper[4840]: I0930 14:56:22.488436 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rzkrh_0ae5abcf-aeb1-4607-8285-cd6ea1d1f912/extract-utilities/0.log" Sep 30 14:56:22 crc kubenswrapper[4840]: I0930 14:56:22.675724 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rzkrh_0ae5abcf-aeb1-4607-8285-cd6ea1d1f912/extract-content/0.log" Sep 30 14:56:22 crc kubenswrapper[4840]: I0930 14:56:22.675758 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rzkrh_0ae5abcf-aeb1-4607-8285-cd6ea1d1f912/extract-utilities/0.log" Sep 30 14:56:22 crc kubenswrapper[4840]: I0930 14:56:22.685088 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rzkrh_0ae5abcf-aeb1-4607-8285-cd6ea1d1f912/extract-content/0.log" Sep 30 14:56:22 crc kubenswrapper[4840]: I0930 14:56:22.847191 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rzkrh_0ae5abcf-aeb1-4607-8285-cd6ea1d1f912/extract-content/0.log" Sep 30 14:56:22 crc kubenswrapper[4840]: I0930 14:56:22.875907 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rzkrh_0ae5abcf-aeb1-4607-8285-cd6ea1d1f912/extract-utilities/0.log" Sep 30 14:56:23 crc kubenswrapper[4840]: I0930 14:56:23.199170 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rzkrh_0ae5abcf-aeb1-4607-8285-cd6ea1d1f912/registry-server/0.log" Sep 30 14:56:51 crc kubenswrapper[4840]: I0930 14:56:51.872805 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:56:51 crc kubenswrapper[4840]: I0930 14:56:51.873425 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:56:51 crc kubenswrapper[4840]: I0930 14:56:51.873477 4840 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-747gk" Sep 30 14:56:51 crc kubenswrapper[4840]: I0930 14:56:51.874065 4840 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f4b62e43937e756d3df8cd25fd431ebd8934b3c7b61115065214e7963ec22cba"} pod="openshift-machine-config-operator/machine-config-daemon-747gk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 14:56:51 crc kubenswrapper[4840]: I0930 14:56:51.874122 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" containerID="cri-o://f4b62e43937e756d3df8cd25fd431ebd8934b3c7b61115065214e7963ec22cba" gracePeriod=600 Sep 30 14:56:52 crc kubenswrapper[4840]: I0930 14:56:52.160680 4840 generic.go:334] "Generic (PLEG): container finished" podID="10e8b890-7f20-4a36-8e03-898620cf599a" containerID="f4b62e43937e756d3df8cd25fd431ebd8934b3c7b61115065214e7963ec22cba" exitCode=0 Sep 30 14:56:52 crc kubenswrapper[4840]: I0930 14:56:52.160751 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" event={"ID":"10e8b890-7f20-4a36-8e03-898620cf599a","Type":"ContainerDied","Data":"f4b62e43937e756d3df8cd25fd431ebd8934b3c7b61115065214e7963ec22cba"} Sep 30 14:56:52 crc kubenswrapper[4840]: I0930 14:56:52.161016 4840 scope.go:117] "RemoveContainer" containerID="bb42e5820a8aaddbc79e0404192d904aa52fd67a0816b292b3605d7cf3a9d5b0" Sep 30 14:56:53 crc kubenswrapper[4840]: I0930 14:56:53.172231 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" event={"ID":"10e8b890-7f20-4a36-8e03-898620cf599a","Type":"ContainerStarted","Data":"409430684b16cc7e5d86eb8dbc5c06065665c2c71a8402382eb805da2eca622c"} Sep 30 14:57:57 crc kubenswrapper[4840]: I0930 14:57:57.494978 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qjl56"] Sep 30 14:57:57 crc kubenswrapper[4840]: E0930 14:57:57.498002 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="510e3570-29ff-4148-8314-a4707f1af37a" containerName="container-00" Sep 30 14:57:57 crc kubenswrapper[4840]: I0930 14:57:57.498022 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="510e3570-29ff-4148-8314-a4707f1af37a" containerName="container-00" Sep 30 14:57:57 crc kubenswrapper[4840]: I0930 14:57:57.498211 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="510e3570-29ff-4148-8314-a4707f1af37a" containerName="container-00" Sep 30 14:57:57 crc kubenswrapper[4840]: I0930 14:57:57.499697 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qjl56" Sep 30 14:57:57 crc kubenswrapper[4840]: I0930 14:57:57.511637 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qjl56"] Sep 30 14:57:57 crc kubenswrapper[4840]: I0930 14:57:57.564409 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65d7b28b-0892-4bec-80e0-76af59dd3996-utilities\") pod \"redhat-marketplace-qjl56\" (UID: \"65d7b28b-0892-4bec-80e0-76af59dd3996\") " pod="openshift-marketplace/redhat-marketplace-qjl56" Sep 30 14:57:57 crc kubenswrapper[4840]: I0930 14:57:57.564520 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zm2vl\" (UniqueName: \"kubernetes.io/projected/65d7b28b-0892-4bec-80e0-76af59dd3996-kube-api-access-zm2vl\") pod \"redhat-marketplace-qjl56\" (UID: \"65d7b28b-0892-4bec-80e0-76af59dd3996\") " pod="openshift-marketplace/redhat-marketplace-qjl56" Sep 30 14:57:57 crc kubenswrapper[4840]: I0930 14:57:57.564627 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65d7b28b-0892-4bec-80e0-76af59dd3996-catalog-content\") pod \"redhat-marketplace-qjl56\" (UID: \"65d7b28b-0892-4bec-80e0-76af59dd3996\") " pod="openshift-marketplace/redhat-marketplace-qjl56" Sep 30 14:57:57 crc kubenswrapper[4840]: I0930 14:57:57.666605 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65d7b28b-0892-4bec-80e0-76af59dd3996-utilities\") pod \"redhat-marketplace-qjl56\" (UID: \"65d7b28b-0892-4bec-80e0-76af59dd3996\") " pod="openshift-marketplace/redhat-marketplace-qjl56" Sep 30 14:57:57 crc kubenswrapper[4840]: I0930 14:57:57.666703 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zm2vl\" (UniqueName: \"kubernetes.io/projected/65d7b28b-0892-4bec-80e0-76af59dd3996-kube-api-access-zm2vl\") pod \"redhat-marketplace-qjl56\" (UID: \"65d7b28b-0892-4bec-80e0-76af59dd3996\") " pod="openshift-marketplace/redhat-marketplace-qjl56" Sep 30 14:57:57 crc kubenswrapper[4840]: I0930 14:57:57.666772 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65d7b28b-0892-4bec-80e0-76af59dd3996-catalog-content\") pod \"redhat-marketplace-qjl56\" (UID: \"65d7b28b-0892-4bec-80e0-76af59dd3996\") " pod="openshift-marketplace/redhat-marketplace-qjl56" Sep 30 14:57:57 crc kubenswrapper[4840]: I0930 14:57:57.667245 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65d7b28b-0892-4bec-80e0-76af59dd3996-utilities\") pod \"redhat-marketplace-qjl56\" (UID: \"65d7b28b-0892-4bec-80e0-76af59dd3996\") " pod="openshift-marketplace/redhat-marketplace-qjl56" Sep 30 14:57:57 crc kubenswrapper[4840]: I0930 14:57:57.667672 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65d7b28b-0892-4bec-80e0-76af59dd3996-catalog-content\") pod \"redhat-marketplace-qjl56\" (UID: \"65d7b28b-0892-4bec-80e0-76af59dd3996\") " pod="openshift-marketplace/redhat-marketplace-qjl56" Sep 30 14:57:57 crc kubenswrapper[4840]: I0930 14:57:57.699213 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zm2vl\" (UniqueName: \"kubernetes.io/projected/65d7b28b-0892-4bec-80e0-76af59dd3996-kube-api-access-zm2vl\") pod \"redhat-marketplace-qjl56\" (UID: \"65d7b28b-0892-4bec-80e0-76af59dd3996\") " pod="openshift-marketplace/redhat-marketplace-qjl56" Sep 30 14:57:57 crc kubenswrapper[4840]: I0930 14:57:57.837440 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qjl56" Sep 30 14:57:58 crc kubenswrapper[4840]: I0930 14:57:58.313772 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qjl56"] Sep 30 14:57:58 crc kubenswrapper[4840]: I0930 14:57:58.792832 4840 generic.go:334] "Generic (PLEG): container finished" podID="65d7b28b-0892-4bec-80e0-76af59dd3996" containerID="8233505cc9efaf8f7d07c8728577599f2d16d2100ed7a3558a70f7f95b116f34" exitCode=0 Sep 30 14:57:58 crc kubenswrapper[4840]: I0930 14:57:58.792890 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qjl56" event={"ID":"65d7b28b-0892-4bec-80e0-76af59dd3996","Type":"ContainerDied","Data":"8233505cc9efaf8f7d07c8728577599f2d16d2100ed7a3558a70f7f95b116f34"} Sep 30 14:57:58 crc kubenswrapper[4840]: I0930 14:57:58.793234 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qjl56" event={"ID":"65d7b28b-0892-4bec-80e0-76af59dd3996","Type":"ContainerStarted","Data":"8d2f7e6d1d7698d7843766742452b2f7d2a5f818453c81b13c4fffd800ab9d36"} Sep 30 14:57:58 crc kubenswrapper[4840]: I0930 14:57:58.797659 4840 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 14:57:59 crc kubenswrapper[4840]: I0930 14:57:59.804541 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qjl56" event={"ID":"65d7b28b-0892-4bec-80e0-76af59dd3996","Type":"ContainerStarted","Data":"fed40f645776a351bb63365f27d305ff33229d8a5ce856f1678998a4e559951f"} Sep 30 14:58:00 crc kubenswrapper[4840]: I0930 14:58:00.817059 4840 generic.go:334] "Generic (PLEG): container finished" podID="65d7b28b-0892-4bec-80e0-76af59dd3996" containerID="fed40f645776a351bb63365f27d305ff33229d8a5ce856f1678998a4e559951f" exitCode=0 Sep 30 14:58:00 crc kubenswrapper[4840]: I0930 14:58:00.817128 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qjl56" event={"ID":"65d7b28b-0892-4bec-80e0-76af59dd3996","Type":"ContainerDied","Data":"fed40f645776a351bb63365f27d305ff33229d8a5ce856f1678998a4e559951f"} Sep 30 14:58:01 crc kubenswrapper[4840]: I0930 14:58:01.828011 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qjl56" event={"ID":"65d7b28b-0892-4bec-80e0-76af59dd3996","Type":"ContainerStarted","Data":"aae5fe952f1d11bdfee1d8487a12a45aac484049e3183568ad3325778b4de732"} Sep 30 14:58:01 crc kubenswrapper[4840]: I0930 14:58:01.849867 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qjl56" podStartSLOduration=2.383235573 podStartE2EDuration="4.849845201s" podCreationTimestamp="2025-09-30 14:57:57 +0000 UTC" firstStartedPulling="2025-09-30 14:57:58.795428289 +0000 UTC m=+3707.424514712" lastFinishedPulling="2025-09-30 14:58:01.262037917 +0000 UTC m=+3709.891124340" observedRunningTime="2025-09-30 14:58:01.841991808 +0000 UTC m=+3710.471078231" watchObservedRunningTime="2025-09-30 14:58:01.849845201 +0000 UTC m=+3710.478931634" Sep 30 14:58:07 crc kubenswrapper[4840]: I0930 14:58:07.837753 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qjl56" Sep 30 14:58:07 crc kubenswrapper[4840]: I0930 14:58:07.838591 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qjl56" Sep 30 14:58:07 crc kubenswrapper[4840]: I0930 14:58:07.907051 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qjl56" Sep 30 14:58:07 crc kubenswrapper[4840]: I0930 14:58:07.957541 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qjl56" Sep 30 14:58:08 crc kubenswrapper[4840]: I0930 14:58:08.149245 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qjl56"] Sep 30 14:58:09 crc kubenswrapper[4840]: I0930 14:58:09.911195 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qjl56" podUID="65d7b28b-0892-4bec-80e0-76af59dd3996" containerName="registry-server" containerID="cri-o://aae5fe952f1d11bdfee1d8487a12a45aac484049e3183568ad3325778b4de732" gracePeriod=2 Sep 30 14:58:10 crc kubenswrapper[4840]: I0930 14:58:10.496003 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qjl56" Sep 30 14:58:10 crc kubenswrapper[4840]: I0930 14:58:10.666869 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zm2vl\" (UniqueName: \"kubernetes.io/projected/65d7b28b-0892-4bec-80e0-76af59dd3996-kube-api-access-zm2vl\") pod \"65d7b28b-0892-4bec-80e0-76af59dd3996\" (UID: \"65d7b28b-0892-4bec-80e0-76af59dd3996\") " Sep 30 14:58:10 crc kubenswrapper[4840]: I0930 14:58:10.667116 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65d7b28b-0892-4bec-80e0-76af59dd3996-catalog-content\") pod \"65d7b28b-0892-4bec-80e0-76af59dd3996\" (UID: \"65d7b28b-0892-4bec-80e0-76af59dd3996\") " Sep 30 14:58:10 crc kubenswrapper[4840]: I0930 14:58:10.667218 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65d7b28b-0892-4bec-80e0-76af59dd3996-utilities\") pod \"65d7b28b-0892-4bec-80e0-76af59dd3996\" (UID: \"65d7b28b-0892-4bec-80e0-76af59dd3996\") " Sep 30 14:58:10 crc kubenswrapper[4840]: I0930 14:58:10.668231 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65d7b28b-0892-4bec-80e0-76af59dd3996-utilities" (OuterVolumeSpecName: "utilities") pod "65d7b28b-0892-4bec-80e0-76af59dd3996" (UID: "65d7b28b-0892-4bec-80e0-76af59dd3996"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:58:10 crc kubenswrapper[4840]: I0930 14:58:10.677157 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65d7b28b-0892-4bec-80e0-76af59dd3996-kube-api-access-zm2vl" (OuterVolumeSpecName: "kube-api-access-zm2vl") pod "65d7b28b-0892-4bec-80e0-76af59dd3996" (UID: "65d7b28b-0892-4bec-80e0-76af59dd3996"). InnerVolumeSpecName "kube-api-access-zm2vl". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:58:10 crc kubenswrapper[4840]: I0930 14:58:10.680101 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65d7b28b-0892-4bec-80e0-76af59dd3996-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "65d7b28b-0892-4bec-80e0-76af59dd3996" (UID: "65d7b28b-0892-4bec-80e0-76af59dd3996"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:58:10 crc kubenswrapper[4840]: I0930 14:58:10.769839 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zm2vl\" (UniqueName: \"kubernetes.io/projected/65d7b28b-0892-4bec-80e0-76af59dd3996-kube-api-access-zm2vl\") on node \"crc\" DevicePath \"\"" Sep 30 14:58:10 crc kubenswrapper[4840]: I0930 14:58:10.769887 4840 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65d7b28b-0892-4bec-80e0-76af59dd3996-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:58:10 crc kubenswrapper[4840]: I0930 14:58:10.769900 4840 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65d7b28b-0892-4bec-80e0-76af59dd3996-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:58:10 crc kubenswrapper[4840]: I0930 14:58:10.922249 4840 generic.go:334] "Generic (PLEG): container finished" podID="65d7b28b-0892-4bec-80e0-76af59dd3996" containerID="aae5fe952f1d11bdfee1d8487a12a45aac484049e3183568ad3325778b4de732" exitCode=0 Sep 30 14:58:10 crc kubenswrapper[4840]: I0930 14:58:10.922292 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qjl56" event={"ID":"65d7b28b-0892-4bec-80e0-76af59dd3996","Type":"ContainerDied","Data":"aae5fe952f1d11bdfee1d8487a12a45aac484049e3183568ad3325778b4de732"} Sep 30 14:58:10 crc kubenswrapper[4840]: I0930 14:58:10.922325 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qjl56" event={"ID":"65d7b28b-0892-4bec-80e0-76af59dd3996","Type":"ContainerDied","Data":"8d2f7e6d1d7698d7843766742452b2f7d2a5f818453c81b13c4fffd800ab9d36"} Sep 30 14:58:10 crc kubenswrapper[4840]: I0930 14:58:10.922341 4840 scope.go:117] "RemoveContainer" containerID="aae5fe952f1d11bdfee1d8487a12a45aac484049e3183568ad3325778b4de732" Sep 30 14:58:10 crc kubenswrapper[4840]: I0930 14:58:10.922398 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qjl56" Sep 30 14:58:10 crc kubenswrapper[4840]: I0930 14:58:10.955766 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qjl56"] Sep 30 14:58:10 crc kubenswrapper[4840]: I0930 14:58:10.959188 4840 scope.go:117] "RemoveContainer" containerID="fed40f645776a351bb63365f27d305ff33229d8a5ce856f1678998a4e559951f" Sep 30 14:58:10 crc kubenswrapper[4840]: I0930 14:58:10.967231 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qjl56"] Sep 30 14:58:10 crc kubenswrapper[4840]: I0930 14:58:10.982201 4840 scope.go:117] "RemoveContainer" containerID="8233505cc9efaf8f7d07c8728577599f2d16d2100ed7a3558a70f7f95b116f34" Sep 30 14:58:11 crc kubenswrapper[4840]: I0930 14:58:11.035645 4840 scope.go:117] "RemoveContainer" containerID="aae5fe952f1d11bdfee1d8487a12a45aac484049e3183568ad3325778b4de732" Sep 30 14:58:11 crc kubenswrapper[4840]: E0930 14:58:11.036506 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aae5fe952f1d11bdfee1d8487a12a45aac484049e3183568ad3325778b4de732\": container with ID starting with aae5fe952f1d11bdfee1d8487a12a45aac484049e3183568ad3325778b4de732 not found: ID does not exist" containerID="aae5fe952f1d11bdfee1d8487a12a45aac484049e3183568ad3325778b4de732" Sep 30 14:58:11 crc kubenswrapper[4840]: I0930 14:58:11.036565 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aae5fe952f1d11bdfee1d8487a12a45aac484049e3183568ad3325778b4de732"} err="failed to get container status \"aae5fe952f1d11bdfee1d8487a12a45aac484049e3183568ad3325778b4de732\": rpc error: code = NotFound desc = could not find container \"aae5fe952f1d11bdfee1d8487a12a45aac484049e3183568ad3325778b4de732\": container with ID starting with aae5fe952f1d11bdfee1d8487a12a45aac484049e3183568ad3325778b4de732 not found: ID does not exist" Sep 30 14:58:11 crc kubenswrapper[4840]: I0930 14:58:11.036592 4840 scope.go:117] "RemoveContainer" containerID="fed40f645776a351bb63365f27d305ff33229d8a5ce856f1678998a4e559951f" Sep 30 14:58:11 crc kubenswrapper[4840]: E0930 14:58:11.037482 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fed40f645776a351bb63365f27d305ff33229d8a5ce856f1678998a4e559951f\": container with ID starting with fed40f645776a351bb63365f27d305ff33229d8a5ce856f1678998a4e559951f not found: ID does not exist" containerID="fed40f645776a351bb63365f27d305ff33229d8a5ce856f1678998a4e559951f" Sep 30 14:58:11 crc kubenswrapper[4840]: I0930 14:58:11.037582 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fed40f645776a351bb63365f27d305ff33229d8a5ce856f1678998a4e559951f"} err="failed to get container status \"fed40f645776a351bb63365f27d305ff33229d8a5ce856f1678998a4e559951f\": rpc error: code = NotFound desc = could not find container \"fed40f645776a351bb63365f27d305ff33229d8a5ce856f1678998a4e559951f\": container with ID starting with fed40f645776a351bb63365f27d305ff33229d8a5ce856f1678998a4e559951f not found: ID does not exist" Sep 30 14:58:11 crc kubenswrapper[4840]: I0930 14:58:11.037604 4840 scope.go:117] "RemoveContainer" containerID="8233505cc9efaf8f7d07c8728577599f2d16d2100ed7a3558a70f7f95b116f34" Sep 30 14:58:11 crc kubenswrapper[4840]: E0930 14:58:11.038018 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8233505cc9efaf8f7d07c8728577599f2d16d2100ed7a3558a70f7f95b116f34\": container with ID starting with 8233505cc9efaf8f7d07c8728577599f2d16d2100ed7a3558a70f7f95b116f34 not found: ID does not exist" containerID="8233505cc9efaf8f7d07c8728577599f2d16d2100ed7a3558a70f7f95b116f34" Sep 30 14:58:11 crc kubenswrapper[4840]: I0930 14:58:11.038057 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8233505cc9efaf8f7d07c8728577599f2d16d2100ed7a3558a70f7f95b116f34"} err="failed to get container status \"8233505cc9efaf8f7d07c8728577599f2d16d2100ed7a3558a70f7f95b116f34\": rpc error: code = NotFound desc = could not find container \"8233505cc9efaf8f7d07c8728577599f2d16d2100ed7a3558a70f7f95b116f34\": container with ID starting with 8233505cc9efaf8f7d07c8728577599f2d16d2100ed7a3558a70f7f95b116f34 not found: ID does not exist" Sep 30 14:58:12 crc kubenswrapper[4840]: I0930 14:58:12.147129 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65d7b28b-0892-4bec-80e0-76af59dd3996" path="/var/lib/kubelet/pods/65d7b28b-0892-4bec-80e0-76af59dd3996/volumes" Sep 30 14:58:20 crc kubenswrapper[4840]: I0930 14:58:20.011479 4840 generic.go:334] "Generic (PLEG): container finished" podID="bcfd9cee-e287-458b-a32a-4f4cf71a2390" containerID="d651f5a0f9580049b759448f8c688a0a6c815f91878f45e22d1ac7a41b9badc2" exitCode=0 Sep 30 14:58:20 crc kubenswrapper[4840]: I0930 14:58:20.011577 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-r4p2h/must-gather-hmmcl" event={"ID":"bcfd9cee-e287-458b-a32a-4f4cf71a2390","Type":"ContainerDied","Data":"d651f5a0f9580049b759448f8c688a0a6c815f91878f45e22d1ac7a41b9badc2"} Sep 30 14:58:20 crc kubenswrapper[4840]: I0930 14:58:20.012583 4840 scope.go:117] "RemoveContainer" containerID="d651f5a0f9580049b759448f8c688a0a6c815f91878f45e22d1ac7a41b9badc2" Sep 30 14:58:20 crc kubenswrapper[4840]: I0930 14:58:20.196705 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-r4p2h_must-gather-hmmcl_bcfd9cee-e287-458b-a32a-4f4cf71a2390/gather/0.log" Sep 30 14:58:28 crc kubenswrapper[4840]: I0930 14:58:28.282092 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-r4p2h/must-gather-hmmcl"] Sep 30 14:58:28 crc kubenswrapper[4840]: I0930 14:58:28.282930 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-r4p2h/must-gather-hmmcl" podUID="bcfd9cee-e287-458b-a32a-4f4cf71a2390" containerName="copy" containerID="cri-o://a881c4baf61c994d460d0ca76822675954ec1bc32aabdc5086385562a9d8b88c" gracePeriod=2 Sep 30 14:58:28 crc kubenswrapper[4840]: I0930 14:58:28.291532 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-r4p2h/must-gather-hmmcl"] Sep 30 14:58:28 crc kubenswrapper[4840]: I0930 14:58:28.765286 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-r4p2h_must-gather-hmmcl_bcfd9cee-e287-458b-a32a-4f4cf71a2390/copy/0.log" Sep 30 14:58:28 crc kubenswrapper[4840]: I0930 14:58:28.765976 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-r4p2h/must-gather-hmmcl" Sep 30 14:58:28 crc kubenswrapper[4840]: I0930 14:58:28.898745 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7wbrx\" (UniqueName: \"kubernetes.io/projected/bcfd9cee-e287-458b-a32a-4f4cf71a2390-kube-api-access-7wbrx\") pod \"bcfd9cee-e287-458b-a32a-4f4cf71a2390\" (UID: \"bcfd9cee-e287-458b-a32a-4f4cf71a2390\") " Sep 30 14:58:28 crc kubenswrapper[4840]: I0930 14:58:28.899041 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bcfd9cee-e287-458b-a32a-4f4cf71a2390-must-gather-output\") pod \"bcfd9cee-e287-458b-a32a-4f4cf71a2390\" (UID: \"bcfd9cee-e287-458b-a32a-4f4cf71a2390\") " Sep 30 14:58:28 crc kubenswrapper[4840]: I0930 14:58:28.905453 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcfd9cee-e287-458b-a32a-4f4cf71a2390-kube-api-access-7wbrx" (OuterVolumeSpecName: "kube-api-access-7wbrx") pod "bcfd9cee-e287-458b-a32a-4f4cf71a2390" (UID: "bcfd9cee-e287-458b-a32a-4f4cf71a2390"). InnerVolumeSpecName "kube-api-access-7wbrx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:58:29 crc kubenswrapper[4840]: I0930 14:58:29.001494 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7wbrx\" (UniqueName: \"kubernetes.io/projected/bcfd9cee-e287-458b-a32a-4f4cf71a2390-kube-api-access-7wbrx\") on node \"crc\" DevicePath \"\"" Sep 30 14:58:29 crc kubenswrapper[4840]: I0930 14:58:29.044508 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bcfd9cee-e287-458b-a32a-4f4cf71a2390-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "bcfd9cee-e287-458b-a32a-4f4cf71a2390" (UID: "bcfd9cee-e287-458b-a32a-4f4cf71a2390"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:58:29 crc kubenswrapper[4840]: I0930 14:58:29.088291 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-r4p2h_must-gather-hmmcl_bcfd9cee-e287-458b-a32a-4f4cf71a2390/copy/0.log" Sep 30 14:58:29 crc kubenswrapper[4840]: I0930 14:58:29.088660 4840 generic.go:334] "Generic (PLEG): container finished" podID="bcfd9cee-e287-458b-a32a-4f4cf71a2390" containerID="a881c4baf61c994d460d0ca76822675954ec1bc32aabdc5086385562a9d8b88c" exitCode=143 Sep 30 14:58:29 crc kubenswrapper[4840]: I0930 14:58:29.088713 4840 scope.go:117] "RemoveContainer" containerID="a881c4baf61c994d460d0ca76822675954ec1bc32aabdc5086385562a9d8b88c" Sep 30 14:58:29 crc kubenswrapper[4840]: I0930 14:58:29.088835 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-r4p2h/must-gather-hmmcl" Sep 30 14:58:29 crc kubenswrapper[4840]: I0930 14:58:29.102707 4840 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bcfd9cee-e287-458b-a32a-4f4cf71a2390-must-gather-output\") on node \"crc\" DevicePath \"\"" Sep 30 14:58:29 crc kubenswrapper[4840]: I0930 14:58:29.108836 4840 scope.go:117] "RemoveContainer" containerID="d651f5a0f9580049b759448f8c688a0a6c815f91878f45e22d1ac7a41b9badc2" Sep 30 14:58:29 crc kubenswrapper[4840]: I0930 14:58:29.178851 4840 scope.go:117] "RemoveContainer" containerID="a881c4baf61c994d460d0ca76822675954ec1bc32aabdc5086385562a9d8b88c" Sep 30 14:58:29 crc kubenswrapper[4840]: E0930 14:58:29.179300 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a881c4baf61c994d460d0ca76822675954ec1bc32aabdc5086385562a9d8b88c\": container with ID starting with a881c4baf61c994d460d0ca76822675954ec1bc32aabdc5086385562a9d8b88c not found: ID does not exist" containerID="a881c4baf61c994d460d0ca76822675954ec1bc32aabdc5086385562a9d8b88c" Sep 30 14:58:29 crc kubenswrapper[4840]: I0930 14:58:29.179332 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a881c4baf61c994d460d0ca76822675954ec1bc32aabdc5086385562a9d8b88c"} err="failed to get container status \"a881c4baf61c994d460d0ca76822675954ec1bc32aabdc5086385562a9d8b88c\": rpc error: code = NotFound desc = could not find container \"a881c4baf61c994d460d0ca76822675954ec1bc32aabdc5086385562a9d8b88c\": container with ID starting with a881c4baf61c994d460d0ca76822675954ec1bc32aabdc5086385562a9d8b88c not found: ID does not exist" Sep 30 14:58:29 crc kubenswrapper[4840]: I0930 14:58:29.179367 4840 scope.go:117] "RemoveContainer" containerID="d651f5a0f9580049b759448f8c688a0a6c815f91878f45e22d1ac7a41b9badc2" Sep 30 14:58:29 crc kubenswrapper[4840]: E0930 14:58:29.179757 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d651f5a0f9580049b759448f8c688a0a6c815f91878f45e22d1ac7a41b9badc2\": container with ID starting with d651f5a0f9580049b759448f8c688a0a6c815f91878f45e22d1ac7a41b9badc2 not found: ID does not exist" containerID="d651f5a0f9580049b759448f8c688a0a6c815f91878f45e22d1ac7a41b9badc2" Sep 30 14:58:29 crc kubenswrapper[4840]: I0930 14:58:29.179939 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d651f5a0f9580049b759448f8c688a0a6c815f91878f45e22d1ac7a41b9badc2"} err="failed to get container status \"d651f5a0f9580049b759448f8c688a0a6c815f91878f45e22d1ac7a41b9badc2\": rpc error: code = NotFound desc = could not find container \"d651f5a0f9580049b759448f8c688a0a6c815f91878f45e22d1ac7a41b9badc2\": container with ID starting with d651f5a0f9580049b759448f8c688a0a6c815f91878f45e22d1ac7a41b9badc2 not found: ID does not exist" Sep 30 14:58:30 crc kubenswrapper[4840]: I0930 14:58:30.131083 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcfd9cee-e287-458b-a32a-4f4cf71a2390" path="/var/lib/kubelet/pods/bcfd9cee-e287-458b-a32a-4f4cf71a2390/volumes" Sep 30 14:59:06 crc kubenswrapper[4840]: I0930 14:59:06.711287 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-gf4gj/must-gather-2hpk7"] Sep 30 14:59:06 crc kubenswrapper[4840]: E0930 14:59:06.718293 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65d7b28b-0892-4bec-80e0-76af59dd3996" containerName="extract-utilities" Sep 30 14:59:06 crc kubenswrapper[4840]: I0930 14:59:06.718318 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="65d7b28b-0892-4bec-80e0-76af59dd3996" containerName="extract-utilities" Sep 30 14:59:06 crc kubenswrapper[4840]: E0930 14:59:06.718338 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcfd9cee-e287-458b-a32a-4f4cf71a2390" containerName="copy" Sep 30 14:59:06 crc kubenswrapper[4840]: I0930 14:59:06.718346 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcfd9cee-e287-458b-a32a-4f4cf71a2390" containerName="copy" Sep 30 14:59:06 crc kubenswrapper[4840]: E0930 14:59:06.718363 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65d7b28b-0892-4bec-80e0-76af59dd3996" containerName="extract-content" Sep 30 14:59:06 crc kubenswrapper[4840]: I0930 14:59:06.718373 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="65d7b28b-0892-4bec-80e0-76af59dd3996" containerName="extract-content" Sep 30 14:59:06 crc kubenswrapper[4840]: E0930 14:59:06.718388 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcfd9cee-e287-458b-a32a-4f4cf71a2390" containerName="gather" Sep 30 14:59:06 crc kubenswrapper[4840]: I0930 14:59:06.718395 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcfd9cee-e287-458b-a32a-4f4cf71a2390" containerName="gather" Sep 30 14:59:06 crc kubenswrapper[4840]: E0930 14:59:06.718410 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65d7b28b-0892-4bec-80e0-76af59dd3996" containerName="registry-server" Sep 30 14:59:06 crc kubenswrapper[4840]: I0930 14:59:06.718417 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="65d7b28b-0892-4bec-80e0-76af59dd3996" containerName="registry-server" Sep 30 14:59:06 crc kubenswrapper[4840]: I0930 14:59:06.718639 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcfd9cee-e287-458b-a32a-4f4cf71a2390" containerName="copy" Sep 30 14:59:06 crc kubenswrapper[4840]: I0930 14:59:06.718664 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="65d7b28b-0892-4bec-80e0-76af59dd3996" containerName="registry-server" Sep 30 14:59:06 crc kubenswrapper[4840]: I0930 14:59:06.718680 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcfd9cee-e287-458b-a32a-4f4cf71a2390" containerName="gather" Sep 30 14:59:06 crc kubenswrapper[4840]: I0930 14:59:06.719684 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-gf4gj/must-gather-2hpk7"] Sep 30 14:59:06 crc kubenswrapper[4840]: I0930 14:59:06.719768 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gf4gj/must-gather-2hpk7" Sep 30 14:59:06 crc kubenswrapper[4840]: I0930 14:59:06.724765 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-gf4gj"/"openshift-service-ca.crt" Sep 30 14:59:06 crc kubenswrapper[4840]: I0930 14:59:06.724987 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-gf4gj"/"kube-root-ca.crt" Sep 30 14:59:06 crc kubenswrapper[4840]: I0930 14:59:06.732531 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-gf4gj"/"default-dockercfg-ntc7d" Sep 30 14:59:06 crc kubenswrapper[4840]: I0930 14:59:06.794249 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/16639026-05f0-4186-9f32-2b9c7e52bff1-must-gather-output\") pod \"must-gather-2hpk7\" (UID: \"16639026-05f0-4186-9f32-2b9c7e52bff1\") " pod="openshift-must-gather-gf4gj/must-gather-2hpk7" Sep 30 14:59:06 crc kubenswrapper[4840]: I0930 14:59:06.794678 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhxbw\" (UniqueName: \"kubernetes.io/projected/16639026-05f0-4186-9f32-2b9c7e52bff1-kube-api-access-fhxbw\") pod \"must-gather-2hpk7\" (UID: \"16639026-05f0-4186-9f32-2b9c7e52bff1\") " pod="openshift-must-gather-gf4gj/must-gather-2hpk7" Sep 30 14:59:06 crc kubenswrapper[4840]: I0930 14:59:06.896316 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhxbw\" (UniqueName: \"kubernetes.io/projected/16639026-05f0-4186-9f32-2b9c7e52bff1-kube-api-access-fhxbw\") pod \"must-gather-2hpk7\" (UID: \"16639026-05f0-4186-9f32-2b9c7e52bff1\") " pod="openshift-must-gather-gf4gj/must-gather-2hpk7" Sep 30 14:59:06 crc kubenswrapper[4840]: I0930 14:59:06.896482 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/16639026-05f0-4186-9f32-2b9c7e52bff1-must-gather-output\") pod \"must-gather-2hpk7\" (UID: \"16639026-05f0-4186-9f32-2b9c7e52bff1\") " pod="openshift-must-gather-gf4gj/must-gather-2hpk7" Sep 30 14:59:06 crc kubenswrapper[4840]: I0930 14:59:06.897108 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/16639026-05f0-4186-9f32-2b9c7e52bff1-must-gather-output\") pod \"must-gather-2hpk7\" (UID: \"16639026-05f0-4186-9f32-2b9c7e52bff1\") " pod="openshift-must-gather-gf4gj/must-gather-2hpk7" Sep 30 14:59:06 crc kubenswrapper[4840]: I0930 14:59:06.917200 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhxbw\" (UniqueName: \"kubernetes.io/projected/16639026-05f0-4186-9f32-2b9c7e52bff1-kube-api-access-fhxbw\") pod \"must-gather-2hpk7\" (UID: \"16639026-05f0-4186-9f32-2b9c7e52bff1\") " pod="openshift-must-gather-gf4gj/must-gather-2hpk7" Sep 30 14:59:07 crc kubenswrapper[4840]: I0930 14:59:07.047793 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gf4gj/must-gather-2hpk7" Sep 30 14:59:08 crc kubenswrapper[4840]: I0930 14:59:07.570177 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-gf4gj/must-gather-2hpk7"] Sep 30 14:59:08 crc kubenswrapper[4840]: I0930 14:59:08.428730 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gf4gj/must-gather-2hpk7" event={"ID":"16639026-05f0-4186-9f32-2b9c7e52bff1","Type":"ContainerStarted","Data":"157e02fdc66df15d3fc935e6af798167c678a29e75c7302f22f037cb92927f7c"} Sep 30 14:59:08 crc kubenswrapper[4840]: I0930 14:59:08.428777 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gf4gj/must-gather-2hpk7" event={"ID":"16639026-05f0-4186-9f32-2b9c7e52bff1","Type":"ContainerStarted","Data":"a46d157e054e258dc85184bcd87f7f9cd75929c160435b559b14641d1c2986c1"} Sep 30 14:59:08 crc kubenswrapper[4840]: I0930 14:59:08.428789 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gf4gj/must-gather-2hpk7" event={"ID":"16639026-05f0-4186-9f32-2b9c7e52bff1","Type":"ContainerStarted","Data":"4665f2d272400e019dd5af88cb7715b9c071456772efa6905a81b56259271f8b"} Sep 30 14:59:08 crc kubenswrapper[4840]: I0930 14:59:08.450536 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-gf4gj/must-gather-2hpk7" podStartSLOduration=2.450518716 podStartE2EDuration="2.450518716s" podCreationTimestamp="2025-09-30 14:59:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:59:08.44216981 +0000 UTC m=+3777.071256233" watchObservedRunningTime="2025-09-30 14:59:08.450518716 +0000 UTC m=+3777.079605139" Sep 30 14:59:11 crc kubenswrapper[4840]: I0930 14:59:11.173490 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-gf4gj/crc-debug-x4l5h"] Sep 30 14:59:11 crc kubenswrapper[4840]: I0930 14:59:11.175213 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gf4gj/crc-debug-x4l5h" Sep 30 14:59:11 crc kubenswrapper[4840]: I0930 14:59:11.294712 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7b5a37b7-2a02-4c6c-9d78-aa178cdf23c2-host\") pod \"crc-debug-x4l5h\" (UID: \"7b5a37b7-2a02-4c6c-9d78-aa178cdf23c2\") " pod="openshift-must-gather-gf4gj/crc-debug-x4l5h" Sep 30 14:59:11 crc kubenswrapper[4840]: I0930 14:59:11.294948 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njd95\" (UniqueName: \"kubernetes.io/projected/7b5a37b7-2a02-4c6c-9d78-aa178cdf23c2-kube-api-access-njd95\") pod \"crc-debug-x4l5h\" (UID: \"7b5a37b7-2a02-4c6c-9d78-aa178cdf23c2\") " pod="openshift-must-gather-gf4gj/crc-debug-x4l5h" Sep 30 14:59:11 crc kubenswrapper[4840]: I0930 14:59:11.396492 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njd95\" (UniqueName: \"kubernetes.io/projected/7b5a37b7-2a02-4c6c-9d78-aa178cdf23c2-kube-api-access-njd95\") pod \"crc-debug-x4l5h\" (UID: \"7b5a37b7-2a02-4c6c-9d78-aa178cdf23c2\") " pod="openshift-must-gather-gf4gj/crc-debug-x4l5h" Sep 30 14:59:11 crc kubenswrapper[4840]: I0930 14:59:11.396578 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7b5a37b7-2a02-4c6c-9d78-aa178cdf23c2-host\") pod \"crc-debug-x4l5h\" (UID: \"7b5a37b7-2a02-4c6c-9d78-aa178cdf23c2\") " pod="openshift-must-gather-gf4gj/crc-debug-x4l5h" Sep 30 14:59:11 crc kubenswrapper[4840]: I0930 14:59:11.396733 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7b5a37b7-2a02-4c6c-9d78-aa178cdf23c2-host\") pod \"crc-debug-x4l5h\" (UID: \"7b5a37b7-2a02-4c6c-9d78-aa178cdf23c2\") " pod="openshift-must-gather-gf4gj/crc-debug-x4l5h" Sep 30 14:59:11 crc kubenswrapper[4840]: I0930 14:59:11.414061 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njd95\" (UniqueName: \"kubernetes.io/projected/7b5a37b7-2a02-4c6c-9d78-aa178cdf23c2-kube-api-access-njd95\") pod \"crc-debug-x4l5h\" (UID: \"7b5a37b7-2a02-4c6c-9d78-aa178cdf23c2\") " pod="openshift-must-gather-gf4gj/crc-debug-x4l5h" Sep 30 14:59:11 crc kubenswrapper[4840]: I0930 14:59:11.497987 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gf4gj/crc-debug-x4l5h" Sep 30 14:59:11 crc kubenswrapper[4840]: W0930 14:59:11.525254 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b5a37b7_2a02_4c6c_9d78_aa178cdf23c2.slice/crio-d8f2dd8bbc11caef452a7422fb1c9605d785e86be2c8cc04000917d688cd70e7 WatchSource:0}: Error finding container d8f2dd8bbc11caef452a7422fb1c9605d785e86be2c8cc04000917d688cd70e7: Status 404 returned error can't find the container with id d8f2dd8bbc11caef452a7422fb1c9605d785e86be2c8cc04000917d688cd70e7 Sep 30 14:59:12 crc kubenswrapper[4840]: I0930 14:59:12.466417 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gf4gj/crc-debug-x4l5h" event={"ID":"7b5a37b7-2a02-4c6c-9d78-aa178cdf23c2","Type":"ContainerStarted","Data":"483b4c8eb835e622b0c96cb2dfbc37f655eb7ab76298701f6bc97c89ee6270f1"} Sep 30 14:59:12 crc kubenswrapper[4840]: I0930 14:59:12.467072 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gf4gj/crc-debug-x4l5h" event={"ID":"7b5a37b7-2a02-4c6c-9d78-aa178cdf23c2","Type":"ContainerStarted","Data":"d8f2dd8bbc11caef452a7422fb1c9605d785e86be2c8cc04000917d688cd70e7"} Sep 30 14:59:12 crc kubenswrapper[4840]: I0930 14:59:12.482000 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-gf4gj/crc-debug-x4l5h" podStartSLOduration=1.481979969 podStartE2EDuration="1.481979969s" podCreationTimestamp="2025-09-30 14:59:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:59:12.481957798 +0000 UTC m=+3781.111044221" watchObservedRunningTime="2025-09-30 14:59:12.481979969 +0000 UTC m=+3781.111066392" Sep 30 14:59:21 crc kubenswrapper[4840]: I0930 14:59:21.871519 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:59:21 crc kubenswrapper[4840]: I0930 14:59:21.872015 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:59:51 crc kubenswrapper[4840]: I0930 14:59:51.871616 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:59:51 crc kubenswrapper[4840]: I0930 14:59:51.872212 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:59:52 crc kubenswrapper[4840]: I0930 14:59:52.064727 4840 scope.go:117] "RemoveContainer" containerID="080b4a392d725b13e94a18caad6a5fb0fd3d342d758f311ad9ec204fa2b1a720" Sep 30 15:00:00 crc kubenswrapper[4840]: I0930 15:00:00.161965 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320740-mw2lk"] Sep 30 15:00:00 crc kubenswrapper[4840]: I0930 15:00:00.163787 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320740-mw2lk" Sep 30 15:00:00 crc kubenswrapper[4840]: I0930 15:00:00.166909 4840 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 30 15:00:00 crc kubenswrapper[4840]: I0930 15:00:00.166915 4840 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 30 15:00:00 crc kubenswrapper[4840]: I0930 15:00:00.187432 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320740-mw2lk"] Sep 30 15:00:00 crc kubenswrapper[4840]: I0930 15:00:00.221027 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d-secret-volume\") pod \"collect-profiles-29320740-mw2lk\" (UID: \"4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320740-mw2lk" Sep 30 15:00:00 crc kubenswrapper[4840]: I0930 15:00:00.221387 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6rqk\" (UniqueName: \"kubernetes.io/projected/4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d-kube-api-access-s6rqk\") pod \"collect-profiles-29320740-mw2lk\" (UID: \"4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320740-mw2lk" Sep 30 15:00:00 crc kubenswrapper[4840]: I0930 15:00:00.221605 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d-config-volume\") pod \"collect-profiles-29320740-mw2lk\" (UID: \"4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320740-mw2lk" Sep 30 15:00:00 crc kubenswrapper[4840]: I0930 15:00:00.324178 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6rqk\" (UniqueName: \"kubernetes.io/projected/4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d-kube-api-access-s6rqk\") pod \"collect-profiles-29320740-mw2lk\" (UID: \"4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320740-mw2lk" Sep 30 15:00:00 crc kubenswrapper[4840]: I0930 15:00:00.324279 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d-config-volume\") pod \"collect-profiles-29320740-mw2lk\" (UID: \"4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320740-mw2lk" Sep 30 15:00:00 crc kubenswrapper[4840]: I0930 15:00:00.324375 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d-secret-volume\") pod \"collect-profiles-29320740-mw2lk\" (UID: \"4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320740-mw2lk" Sep 30 15:00:00 crc kubenswrapper[4840]: I0930 15:00:00.325348 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d-config-volume\") pod \"collect-profiles-29320740-mw2lk\" (UID: \"4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320740-mw2lk" Sep 30 15:00:00 crc kubenswrapper[4840]: I0930 15:00:00.332283 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d-secret-volume\") pod \"collect-profiles-29320740-mw2lk\" (UID: \"4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320740-mw2lk" Sep 30 15:00:00 crc kubenswrapper[4840]: I0930 15:00:00.345313 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6rqk\" (UniqueName: \"kubernetes.io/projected/4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d-kube-api-access-s6rqk\") pod \"collect-profiles-29320740-mw2lk\" (UID: \"4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320740-mw2lk" Sep 30 15:00:00 crc kubenswrapper[4840]: I0930 15:00:00.507792 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320740-mw2lk" Sep 30 15:00:00 crc kubenswrapper[4840]: I0930 15:00:00.946619 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320740-mw2lk"] Sep 30 15:00:01 crc kubenswrapper[4840]: I0930 15:00:01.886282 4840 generic.go:334] "Generic (PLEG): container finished" podID="4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d" containerID="0f74072fcc9c70bee4562dce847fbde459ab899941963cd2d99205a83aadd017" exitCode=0 Sep 30 15:00:01 crc kubenswrapper[4840]: I0930 15:00:01.886342 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320740-mw2lk" event={"ID":"4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d","Type":"ContainerDied","Data":"0f74072fcc9c70bee4562dce847fbde459ab899941963cd2d99205a83aadd017"} Sep 30 15:00:01 crc kubenswrapper[4840]: I0930 15:00:01.886572 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320740-mw2lk" event={"ID":"4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d","Type":"ContainerStarted","Data":"282da7e90a72818a78d3455e9172f8f007f701bb1e03a1f12a5f16945d74b9de"} Sep 30 15:00:03 crc kubenswrapper[4840]: I0930 15:00:03.253113 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320740-mw2lk" Sep 30 15:00:03 crc kubenswrapper[4840]: I0930 15:00:03.282512 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d-config-volume\") pod \"4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d\" (UID: \"4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d\") " Sep 30 15:00:03 crc kubenswrapper[4840]: I0930 15:00:03.282758 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d-secret-volume\") pod \"4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d\" (UID: \"4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d\") " Sep 30 15:00:03 crc kubenswrapper[4840]: I0930 15:00:03.282846 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s6rqk\" (UniqueName: \"kubernetes.io/projected/4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d-kube-api-access-s6rqk\") pod \"4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d\" (UID: \"4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d\") " Sep 30 15:00:03 crc kubenswrapper[4840]: I0930 15:00:03.283818 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d-config-volume" (OuterVolumeSpecName: "config-volume") pod "4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d" (UID: "4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 15:00:03 crc kubenswrapper[4840]: I0930 15:00:03.296102 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d" (UID: "4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 15:00:03 crc kubenswrapper[4840]: I0930 15:00:03.296239 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d-kube-api-access-s6rqk" (OuterVolumeSpecName: "kube-api-access-s6rqk") pod "4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d" (UID: "4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d"). InnerVolumeSpecName "kube-api-access-s6rqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:00:03 crc kubenswrapper[4840]: I0930 15:00:03.385656 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s6rqk\" (UniqueName: \"kubernetes.io/projected/4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d-kube-api-access-s6rqk\") on node \"crc\" DevicePath \"\"" Sep 30 15:00:03 crc kubenswrapper[4840]: I0930 15:00:03.386089 4840 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d-config-volume\") on node \"crc\" DevicePath \"\"" Sep 30 15:00:03 crc kubenswrapper[4840]: I0930 15:00:03.386103 4840 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 30 15:00:03 crc kubenswrapper[4840]: I0930 15:00:03.905897 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320740-mw2lk" event={"ID":"4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d","Type":"ContainerDied","Data":"282da7e90a72818a78d3455e9172f8f007f701bb1e03a1f12a5f16945d74b9de"} Sep 30 15:00:03 crc kubenswrapper[4840]: I0930 15:00:03.905936 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320740-mw2lk" Sep 30 15:00:03 crc kubenswrapper[4840]: I0930 15:00:03.905953 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="282da7e90a72818a78d3455e9172f8f007f701bb1e03a1f12a5f16945d74b9de" Sep 30 15:00:04 crc kubenswrapper[4840]: I0930 15:00:04.330892 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320695-ml6k2"] Sep 30 15:00:04 crc kubenswrapper[4840]: I0930 15:00:04.339176 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320695-ml6k2"] Sep 30 15:00:06 crc kubenswrapper[4840]: I0930 15:00:06.128582 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af8e3cc6-9628-4d6b-81c6-41a55cfb8935" path="/var/lib/kubelet/pods/af8e3cc6-9628-4d6b-81c6-41a55cfb8935/volumes" Sep 30 15:00:10 crc kubenswrapper[4840]: I0930 15:00:10.540410 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-784f485bd4-gztrl_f7f068af-286b-4d75-ace9-247f0432de8a/barbican-api/0.log" Sep 30 15:00:10 crc kubenswrapper[4840]: I0930 15:00:10.550701 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-784f485bd4-gztrl_f7f068af-286b-4d75-ace9-247f0432de8a/barbican-api-log/0.log" Sep 30 15:00:10 crc kubenswrapper[4840]: I0930 15:00:10.722201 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-67bf9f4b94-nk9h8_ae9a65ac-f1a7-4ef0-aeba-aca57985ae87/barbican-keystone-listener/0.log" Sep 30 15:00:10 crc kubenswrapper[4840]: I0930 15:00:10.753629 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-67bf9f4b94-nk9h8_ae9a65ac-f1a7-4ef0-aeba-aca57985ae87/barbican-keystone-listener-log/0.log" Sep 30 15:00:10 crc kubenswrapper[4840]: I0930 15:00:10.884462 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-668f8fbd45-fpqbx_0fa943f1-8698-41e3-a893-fd1c8845076a/barbican-worker/0.log" Sep 30 15:00:10 crc kubenswrapper[4840]: I0930 15:00:10.939390 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-668f8fbd45-fpqbx_0fa943f1-8698-41e3-a893-fd1c8845076a/barbican-worker-log/0.log" Sep 30 15:00:11 crc kubenswrapper[4840]: I0930 15:00:11.164296 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-6ntwv_558c75bb-7947-423d-b3aa-1b2ab4ba2d1d/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:00:11 crc kubenswrapper[4840]: I0930 15:00:11.344728 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_5673b7b1-d1e1-47ac-8a65-c210430ac4a5/ceilometer-central-agent/0.log" Sep 30 15:00:11 crc kubenswrapper[4840]: I0930 15:00:11.369705 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_5673b7b1-d1e1-47ac-8a65-c210430ac4a5/proxy-httpd/0.log" Sep 30 15:00:11 crc kubenswrapper[4840]: I0930 15:00:11.383388 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_5673b7b1-d1e1-47ac-8a65-c210430ac4a5/ceilometer-notification-agent/0.log" Sep 30 15:00:11 crc kubenswrapper[4840]: I0930 15:00:11.521273 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_5673b7b1-d1e1-47ac-8a65-c210430ac4a5/sg-core/0.log" Sep 30 15:00:11 crc kubenswrapper[4840]: I0930 15:00:11.623933 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_eb01e3a9-61de-4dc8-954b-221d80a675fb/cinder-api/0.log" Sep 30 15:00:11 crc kubenswrapper[4840]: I0930 15:00:11.699904 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_eb01e3a9-61de-4dc8-954b-221d80a675fb/cinder-api-log/0.log" Sep 30 15:00:11 crc kubenswrapper[4840]: I0930 15:00:11.880801 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_01feefa5-d23e-4189-9d57-84bae2e247e2/cinder-scheduler/0.log" Sep 30 15:00:11 crc kubenswrapper[4840]: I0930 15:00:11.920238 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_01feefa5-d23e-4189-9d57-84bae2e247e2/probe/0.log" Sep 30 15:00:12 crc kubenswrapper[4840]: I0930 15:00:12.134406 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-xvhf8_b3a0bf4c-6d14-49c4-9a98-3506318ee798/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:00:12 crc kubenswrapper[4840]: I0930 15:00:12.241403 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-9p2lj_cbc06ec4-8aae-4192-a201-d2b0d13b29b1/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:00:12 crc kubenswrapper[4840]: I0930 15:00:12.339006 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-bmz4g_797b0ada-6af8-4509-8a81-6b97c263fdc3/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:00:12 crc kubenswrapper[4840]: I0930 15:00:12.542467 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-87k2l_462dd9f7-ec69-4883-aa78-9b014c449baa/init/0.log" Sep 30 15:00:12 crc kubenswrapper[4840]: I0930 15:00:12.740720 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-87k2l_462dd9f7-ec69-4883-aa78-9b014c449baa/init/0.log" Sep 30 15:00:12 crc kubenswrapper[4840]: I0930 15:00:12.791544 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-87k2l_462dd9f7-ec69-4883-aa78-9b014c449baa/dnsmasq-dns/0.log" Sep 30 15:00:12 crc kubenswrapper[4840]: I0930 15:00:12.984683 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-cbh6n_17f4c34f-39a2-44a1-9ce0-d5a4bf1f9780/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:00:12 crc kubenswrapper[4840]: I0930 15:00:12.997531 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_b1e73274-948e-4813-946c-0d98ce327e91/glance-httpd/0.log" Sep 30 15:00:13 crc kubenswrapper[4840]: I0930 15:00:13.152380 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_b1e73274-948e-4813-946c-0d98ce327e91/glance-log/0.log" Sep 30 15:00:13 crc kubenswrapper[4840]: I0930 15:00:13.237536 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_c58eeb6f-560e-4abf-b604-65a3d2819ac5/glance-httpd/0.log" Sep 30 15:00:13 crc kubenswrapper[4840]: I0930 15:00:13.350592 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_c58eeb6f-560e-4abf-b604-65a3d2819ac5/glance-log/0.log" Sep 30 15:00:13 crc kubenswrapper[4840]: I0930 15:00:13.503838 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-94dfcc48d-mlrjd_1138b51b-edf9-42d2-8ea4-4d4b9122eae4/horizon/0.log" Sep 30 15:00:13 crc kubenswrapper[4840]: I0930 15:00:13.694482 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-nwgmx_cd627175-3b3b-4a69-9f1c-f64828267f40/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:00:13 crc kubenswrapper[4840]: I0930 15:00:13.911264 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-hw58j_3c378bd2-1562-466e-aafb-7a597487f2c0/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:00:13 crc kubenswrapper[4840]: I0930 15:00:13.942471 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-94dfcc48d-mlrjd_1138b51b-edf9-42d2-8ea4-4d4b9122eae4/horizon-log/0.log" Sep 30 15:00:14 crc kubenswrapper[4840]: I0930 15:00:14.169229 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_26d4f9e8-9d18-45ec-8117-39c0714ce1b0/kube-state-metrics/0.log" Sep 30 15:00:14 crc kubenswrapper[4840]: I0930 15:00:14.180310 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6c78d49dc8-kqrxd_692fd2c9-5168-48fb-9909-7b1cab678b75/keystone-api/0.log" Sep 30 15:00:14 crc kubenswrapper[4840]: I0930 15:00:14.385160 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-m584b_9f66833a-f0de-410b-9459-80203d4291a6/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:00:14 crc kubenswrapper[4840]: I0930 15:00:14.756308 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-658bb978b7-lrh6p_e8e55cd0-dac0-402d-beb7-6482894554fb/neutron-httpd/0.log" Sep 30 15:00:14 crc kubenswrapper[4840]: I0930 15:00:14.761728 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-658bb978b7-lrh6p_e8e55cd0-dac0-402d-beb7-6482894554fb/neutron-api/0.log" Sep 30 15:00:14 crc kubenswrapper[4840]: I0930 15:00:14.916327 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-tkrm8_800dcd1d-c19d-4841-b3ab-81c312d8a754/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:00:15 crc kubenswrapper[4840]: I0930 15:00:15.539130 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_eade78f5-391f-455f-b61c-4300f8d9f53c/nova-api-log/0.log" Sep 30 15:00:15 crc kubenswrapper[4840]: I0930 15:00:15.722446 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_739533f4-30e5-4941-a7f3-d43a945cd8e8/nova-cell0-conductor-conductor/0.log" Sep 30 15:00:15 crc kubenswrapper[4840]: I0930 15:00:15.832037 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_eade78f5-391f-455f-b61c-4300f8d9f53c/nova-api-api/0.log" Sep 30 15:00:16 crc kubenswrapper[4840]: I0930 15:00:16.080703 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_b63a8a4d-0f06-4f78-9f54-9b4e81183705/nova-cell1-conductor-conductor/0.log" Sep 30 15:00:16 crc kubenswrapper[4840]: I0930 15:00:16.219955 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_194d271a-568f-4295-84f2-80cf46e1f123/nova-cell1-novncproxy-novncproxy/0.log" Sep 30 15:00:16 crc kubenswrapper[4840]: I0930 15:00:16.380201 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-5q27c_20b6236d-9397-4095-a35a-6c24222b852f/nova-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:00:16 crc kubenswrapper[4840]: I0930 15:00:16.601165 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_0f1f889e-0a43-436f-adfe-6bb0a8e9b909/nova-metadata-log/0.log" Sep 30 15:00:17 crc kubenswrapper[4840]: I0930 15:00:17.076880 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4w8xk"] Sep 30 15:00:17 crc kubenswrapper[4840]: E0930 15:00:17.077418 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d" containerName="collect-profiles" Sep 30 15:00:17 crc kubenswrapper[4840]: I0930 15:00:17.077441 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d" containerName="collect-profiles" Sep 30 15:00:17 crc kubenswrapper[4840]: I0930 15:00:17.077513 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_ca7038c5-867f-4546-b846-7a6beae1fa1f/nova-scheduler-scheduler/0.log" Sep 30 15:00:17 crc kubenswrapper[4840]: I0930 15:00:17.077737 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ef0a70d-f38c-4e33-bd79-54a9a0e1ae4d" containerName="collect-profiles" Sep 30 15:00:17 crc kubenswrapper[4840]: I0930 15:00:17.079502 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4w8xk" Sep 30 15:00:17 crc kubenswrapper[4840]: I0930 15:00:17.102844 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4w8xk"] Sep 30 15:00:17 crc kubenswrapper[4840]: I0930 15:00:17.139126 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvr4r\" (UniqueName: \"kubernetes.io/projected/518360cb-03c8-455f-8113-7c3a0493ff7c-kube-api-access-qvr4r\") pod \"redhat-operators-4w8xk\" (UID: \"518360cb-03c8-455f-8113-7c3a0493ff7c\") " pod="openshift-marketplace/redhat-operators-4w8xk" Sep 30 15:00:17 crc kubenswrapper[4840]: I0930 15:00:17.139184 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/518360cb-03c8-455f-8113-7c3a0493ff7c-catalog-content\") pod \"redhat-operators-4w8xk\" (UID: \"518360cb-03c8-455f-8113-7c3a0493ff7c\") " pod="openshift-marketplace/redhat-operators-4w8xk" Sep 30 15:00:17 crc kubenswrapper[4840]: I0930 15:00:17.139417 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/518360cb-03c8-455f-8113-7c3a0493ff7c-utilities\") pod \"redhat-operators-4w8xk\" (UID: \"518360cb-03c8-455f-8113-7c3a0493ff7c\") " pod="openshift-marketplace/redhat-operators-4w8xk" Sep 30 15:00:17 crc kubenswrapper[4840]: I0930 15:00:17.240304 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/518360cb-03c8-455f-8113-7c3a0493ff7c-utilities\") pod \"redhat-operators-4w8xk\" (UID: \"518360cb-03c8-455f-8113-7c3a0493ff7c\") " pod="openshift-marketplace/redhat-operators-4w8xk" Sep 30 15:00:17 crc kubenswrapper[4840]: I0930 15:00:17.240663 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvr4r\" (UniqueName: \"kubernetes.io/projected/518360cb-03c8-455f-8113-7c3a0493ff7c-kube-api-access-qvr4r\") pod \"redhat-operators-4w8xk\" (UID: \"518360cb-03c8-455f-8113-7c3a0493ff7c\") " pod="openshift-marketplace/redhat-operators-4w8xk" Sep 30 15:00:17 crc kubenswrapper[4840]: I0930 15:00:17.240777 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/518360cb-03c8-455f-8113-7c3a0493ff7c-catalog-content\") pod \"redhat-operators-4w8xk\" (UID: \"518360cb-03c8-455f-8113-7c3a0493ff7c\") " pod="openshift-marketplace/redhat-operators-4w8xk" Sep 30 15:00:17 crc kubenswrapper[4840]: I0930 15:00:17.241498 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/518360cb-03c8-455f-8113-7c3a0493ff7c-catalog-content\") pod \"redhat-operators-4w8xk\" (UID: \"518360cb-03c8-455f-8113-7c3a0493ff7c\") " pod="openshift-marketplace/redhat-operators-4w8xk" Sep 30 15:00:17 crc kubenswrapper[4840]: I0930 15:00:17.241948 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/518360cb-03c8-455f-8113-7c3a0493ff7c-utilities\") pod \"redhat-operators-4w8xk\" (UID: \"518360cb-03c8-455f-8113-7c3a0493ff7c\") " pod="openshift-marketplace/redhat-operators-4w8xk" Sep 30 15:00:17 crc kubenswrapper[4840]: I0930 15:00:17.277871 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvr4r\" (UniqueName: \"kubernetes.io/projected/518360cb-03c8-455f-8113-7c3a0493ff7c-kube-api-access-qvr4r\") pod \"redhat-operators-4w8xk\" (UID: \"518360cb-03c8-455f-8113-7c3a0493ff7c\") " pod="openshift-marketplace/redhat-operators-4w8xk" Sep 30 15:00:17 crc kubenswrapper[4840]: I0930 15:00:17.386882 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_41b0d67b-4ef4-4ebb-b000-b37bbdac6b34/mysql-bootstrap/0.log" Sep 30 15:00:17 crc kubenswrapper[4840]: I0930 15:00:17.430679 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4w8xk" Sep 30 15:00:17 crc kubenswrapper[4840]: I0930 15:00:17.667967 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_41b0d67b-4ef4-4ebb-b000-b37bbdac6b34/mysql-bootstrap/0.log" Sep 30 15:00:17 crc kubenswrapper[4840]: I0930 15:00:17.706649 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_41b0d67b-4ef4-4ebb-b000-b37bbdac6b34/galera/0.log" Sep 30 15:00:17 crc kubenswrapper[4840]: I0930 15:00:17.984964 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4w8xk"] Sep 30 15:00:18 crc kubenswrapper[4840]: I0930 15:00:18.005243 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f/mysql-bootstrap/0.log" Sep 30 15:00:18 crc kubenswrapper[4840]: I0930 15:00:18.043843 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4w8xk" event={"ID":"518360cb-03c8-455f-8113-7c3a0493ff7c","Type":"ContainerStarted","Data":"6cc60012f998341c5911c625e3aa767d3dc7e2e376aabc1e7ac2b45eeea78fe7"} Sep 30 15:00:18 crc kubenswrapper[4840]: I0930 15:00:18.266881 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_0f1f889e-0a43-436f-adfe-6bb0a8e9b909/nova-metadata-metadata/0.log" Sep 30 15:00:18 crc kubenswrapper[4840]: I0930 15:00:18.322371 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f/mysql-bootstrap/0.log" Sep 30 15:00:18 crc kubenswrapper[4840]: I0930 15:00:18.425621 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_9eddd8f2-9e3b-43cd-ab8f-0d5c36b2dd3f/galera/0.log" Sep 30 15:00:18 crc kubenswrapper[4840]: I0930 15:00:18.681034 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_e234012b-317e-4c79-81a3-3991e1296058/openstackclient/0.log" Sep 30 15:00:18 crc kubenswrapper[4840]: I0930 15:00:18.967065 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-cngvz_e8a63b50-ecd5-4993-a890-8c94bc2d5e60/ovn-controller/0.log" Sep 30 15:00:18 crc kubenswrapper[4840]: I0930 15:00:18.977659 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-2bdgj_a789c20a-6871-4866-bfa2-c69efd876afb/openstack-network-exporter/0.log" Sep 30 15:00:19 crc kubenswrapper[4840]: I0930 15:00:19.054759 4840 generic.go:334] "Generic (PLEG): container finished" podID="518360cb-03c8-455f-8113-7c3a0493ff7c" containerID="11e1516b31d5532eb60ffe13ceb9b893ebd6226cb5c0b72f908537b1f8e71ad3" exitCode=0 Sep 30 15:00:19 crc kubenswrapper[4840]: I0930 15:00:19.054840 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4w8xk" event={"ID":"518360cb-03c8-455f-8113-7c3a0493ff7c","Type":"ContainerDied","Data":"11e1516b31d5532eb60ffe13ceb9b893ebd6226cb5c0b72f908537b1f8e71ad3"} Sep 30 15:00:19 crc kubenswrapper[4840]: I0930 15:00:19.232424 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fc5jg_7d1f7d84-d053-4c95-83ac-ecbff773463d/ovsdb-server-init/0.log" Sep 30 15:00:19 crc kubenswrapper[4840]: I0930 15:00:19.430452 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fc5jg_7d1f7d84-d053-4c95-83ac-ecbff773463d/ovsdb-server-init/0.log" Sep 30 15:00:19 crc kubenswrapper[4840]: I0930 15:00:19.435872 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fc5jg_7d1f7d84-d053-4c95-83ac-ecbff773463d/ovs-vswitchd/0.log" Sep 30 15:00:19 crc kubenswrapper[4840]: I0930 15:00:19.439305 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fc5jg_7d1f7d84-d053-4c95-83ac-ecbff773463d/ovsdb-server/0.log" Sep 30 15:00:19 crc kubenswrapper[4840]: I0930 15:00:19.735353 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-tsm4h_bc398435-386f-4d82-8ba5-618d3795223e/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:00:20 crc kubenswrapper[4840]: I0930 15:00:20.043055 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_2338a6b1-3726-4994-9325-ee16f93e28c0/openstack-network-exporter/0.log" Sep 30 15:00:20 crc kubenswrapper[4840]: I0930 15:00:20.079014 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4w8xk" event={"ID":"518360cb-03c8-455f-8113-7c3a0493ff7c","Type":"ContainerStarted","Data":"94e5c91d166f6a6ab7e3247013f8e9bd52bad848b93f42b2c0817e2ad4ba9d2e"} Sep 30 15:00:20 crc kubenswrapper[4840]: I0930 15:00:20.107640 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_2338a6b1-3726-4994-9325-ee16f93e28c0/ovn-northd/0.log" Sep 30 15:00:20 crc kubenswrapper[4840]: I0930 15:00:20.268071 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_2082b33d-3b6c-449a-9bdd-63d665a3d313/openstack-network-exporter/0.log" Sep 30 15:00:20 crc kubenswrapper[4840]: I0930 15:00:20.408024 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_2082b33d-3b6c-449a-9bdd-63d665a3d313/ovsdbserver-nb/0.log" Sep 30 15:00:20 crc kubenswrapper[4840]: I0930 15:00:20.514992 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_93d37b01-5288-4d31-84db-3fd7e7840c3b/openstack-network-exporter/0.log" Sep 30 15:00:20 crc kubenswrapper[4840]: I0930 15:00:20.617284 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_93d37b01-5288-4d31-84db-3fd7e7840c3b/ovsdbserver-sb/0.log" Sep 30 15:00:20 crc kubenswrapper[4840]: I0930 15:00:20.871455 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6cbd5fdcfb-hn4pn_c1eaca45-6337-4f35-ac67-dfa0fc83e1bb/placement-api/0.log" Sep 30 15:00:20 crc kubenswrapper[4840]: I0930 15:00:20.930369 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6cbd5fdcfb-hn4pn_c1eaca45-6337-4f35-ac67-dfa0fc83e1bb/placement-log/0.log" Sep 30 15:00:21 crc kubenswrapper[4840]: I0930 15:00:21.113930 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_533b8943-b9ea-4062-9dd2-ed1b29d4f2c6/setup-container/0.log" Sep 30 15:00:21 crc kubenswrapper[4840]: I0930 15:00:21.293514 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_533b8943-b9ea-4062-9dd2-ed1b29d4f2c6/rabbitmq/0.log" Sep 30 15:00:21 crc kubenswrapper[4840]: I0930 15:00:21.311283 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_533b8943-b9ea-4062-9dd2-ed1b29d4f2c6/setup-container/0.log" Sep 30 15:00:21 crc kubenswrapper[4840]: I0930 15:00:21.495801 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_97b46089-3117-4a2a-b9db-ada9d0daca8b/setup-container/0.log" Sep 30 15:00:21 crc kubenswrapper[4840]: I0930 15:00:21.681342 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_97b46089-3117-4a2a-b9db-ada9d0daca8b/rabbitmq/0.log" Sep 30 15:00:21 crc kubenswrapper[4840]: I0930 15:00:21.735094 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_97b46089-3117-4a2a-b9db-ada9d0daca8b/setup-container/0.log" Sep 30 15:00:21 crc kubenswrapper[4840]: I0930 15:00:21.871274 4840 patch_prober.go:28] interesting pod/machine-config-daemon-747gk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 15:00:21 crc kubenswrapper[4840]: I0930 15:00:21.871589 4840 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 15:00:21 crc kubenswrapper[4840]: I0930 15:00:21.871656 4840 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-747gk" Sep 30 15:00:21 crc kubenswrapper[4840]: I0930 15:00:21.872410 4840 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"409430684b16cc7e5d86eb8dbc5c06065665c2c71a8402382eb805da2eca622c"} pod="openshift-machine-config-operator/machine-config-daemon-747gk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 15:00:21 crc kubenswrapper[4840]: I0930 15:00:21.872471 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" containerName="machine-config-daemon" containerID="cri-o://409430684b16cc7e5d86eb8dbc5c06065665c2c71a8402382eb805da2eca622c" gracePeriod=600 Sep 30 15:00:22 crc kubenswrapper[4840]: I0930 15:00:22.029023 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-xkx5r_86b7d072-1a56-454f-ad5f-a098a0977122/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:00:22 crc kubenswrapper[4840]: I0930 15:00:22.099924 4840 generic.go:334] "Generic (PLEG): container finished" podID="10e8b890-7f20-4a36-8e03-898620cf599a" containerID="409430684b16cc7e5d86eb8dbc5c06065665c2c71a8402382eb805da2eca622c" exitCode=0 Sep 30 15:00:22 crc kubenswrapper[4840]: I0930 15:00:22.099969 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" event={"ID":"10e8b890-7f20-4a36-8e03-898620cf599a","Type":"ContainerDied","Data":"409430684b16cc7e5d86eb8dbc5c06065665c2c71a8402382eb805da2eca622c"} Sep 30 15:00:22 crc kubenswrapper[4840]: I0930 15:00:22.100005 4840 scope.go:117] "RemoveContainer" containerID="f4b62e43937e756d3df8cd25fd431ebd8934b3c7b61115065214e7963ec22cba" Sep 30 15:00:22 crc kubenswrapper[4840]: I0930 15:00:22.116395 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-gtxqx_84f24beb-97af-44e4-b529-41138e5414ad/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:00:22 crc kubenswrapper[4840]: I0930 15:00:22.291881 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-q5t88_c1073e6a-94e1-4fad-a116-ff4e145a0d22/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:00:22 crc kubenswrapper[4840]: E0930 15:00:22.298111 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 15:00:22 crc kubenswrapper[4840]: I0930 15:00:22.448379 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-nn4ls_a7b69c13-1a32-4676-9912-e117b0b59848/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:00:22 crc kubenswrapper[4840]: I0930 15:00:22.704155 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-dbvwh_37a41ba6-72d7-40d8-a305-153b68f100a8/ssh-known-hosts-edpm-deployment/0.log" Sep 30 15:00:22 crc kubenswrapper[4840]: I0930 15:00:22.908929 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5d586fc9-9z92d_84813c70-f94c-4556-a498-10e17fb354f1/proxy-httpd/0.log" Sep 30 15:00:23 crc kubenswrapper[4840]: I0930 15:00:23.006433 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5d586fc9-9z92d_84813c70-f94c-4556-a498-10e17fb354f1/proxy-server/0.log" Sep 30 15:00:23 crc kubenswrapper[4840]: I0930 15:00:23.127042 4840 scope.go:117] "RemoveContainer" containerID="409430684b16cc7e5d86eb8dbc5c06065665c2c71a8402382eb805da2eca622c" Sep 30 15:00:23 crc kubenswrapper[4840]: E0930 15:00:23.127771 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 15:00:23 crc kubenswrapper[4840]: I0930 15:00:23.167677 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-txwn9_d8412105-132d-4656-8fd8-296cbd8a1ace/swift-ring-rebalance/0.log" Sep 30 15:00:23 crc kubenswrapper[4840]: I0930 15:00:23.267330 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66a7a4e6-dd49-4589-bd88-2f5fdf0af70f/account-auditor/0.log" Sep 30 15:00:23 crc kubenswrapper[4840]: I0930 15:00:23.379534 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66a7a4e6-dd49-4589-bd88-2f5fdf0af70f/account-reaper/0.log" Sep 30 15:00:23 crc kubenswrapper[4840]: I0930 15:00:23.493020 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66a7a4e6-dd49-4589-bd88-2f5fdf0af70f/account-replicator/0.log" Sep 30 15:00:23 crc kubenswrapper[4840]: I0930 15:00:23.568052 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66a7a4e6-dd49-4589-bd88-2f5fdf0af70f/container-auditor/0.log" Sep 30 15:00:23 crc kubenswrapper[4840]: I0930 15:00:23.585028 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66a7a4e6-dd49-4589-bd88-2f5fdf0af70f/account-server/0.log" Sep 30 15:00:23 crc kubenswrapper[4840]: I0930 15:00:23.735741 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66a7a4e6-dd49-4589-bd88-2f5fdf0af70f/container-replicator/0.log" Sep 30 15:00:23 crc kubenswrapper[4840]: I0930 15:00:23.789091 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66a7a4e6-dd49-4589-bd88-2f5fdf0af70f/container-server/0.log" Sep 30 15:00:23 crc kubenswrapper[4840]: I0930 15:00:23.813912 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66a7a4e6-dd49-4589-bd88-2f5fdf0af70f/container-updater/0.log" Sep 30 15:00:23 crc kubenswrapper[4840]: I0930 15:00:23.987819 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66a7a4e6-dd49-4589-bd88-2f5fdf0af70f/object-expirer/0.log" Sep 30 15:00:24 crc kubenswrapper[4840]: I0930 15:00:24.020137 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66a7a4e6-dd49-4589-bd88-2f5fdf0af70f/object-auditor/0.log" Sep 30 15:00:24 crc kubenswrapper[4840]: I0930 15:00:24.117622 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66a7a4e6-dd49-4589-bd88-2f5fdf0af70f/object-replicator/0.log" Sep 30 15:00:24 crc kubenswrapper[4840]: I0930 15:00:24.166261 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66a7a4e6-dd49-4589-bd88-2f5fdf0af70f/object-server/0.log" Sep 30 15:00:24 crc kubenswrapper[4840]: I0930 15:00:24.212842 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66a7a4e6-dd49-4589-bd88-2f5fdf0af70f/object-updater/0.log" Sep 30 15:00:24 crc kubenswrapper[4840]: I0930 15:00:24.323232 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66a7a4e6-dd49-4589-bd88-2f5fdf0af70f/rsync/0.log" Sep 30 15:00:24 crc kubenswrapper[4840]: I0930 15:00:24.394138 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66a7a4e6-dd49-4589-bd88-2f5fdf0af70f/swift-recon-cron/0.log" Sep 30 15:00:24 crc kubenswrapper[4840]: I0930 15:00:24.589454 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-dzddj_73a8aa6c-814e-420f-bb1d-666259304a7e/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:00:24 crc kubenswrapper[4840]: I0930 15:00:24.703195 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_1e0776f6-f8bf-4ef7-a68b-a48fb2e5f545/tempest-tests-tempest-tests-runner/0.log" Sep 30 15:00:24 crc kubenswrapper[4840]: I0930 15:00:24.835382 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_caaa008d-dbdc-499e-9b2a-95c99517e733/test-operator-logs-container/0.log" Sep 30 15:00:25 crc kubenswrapper[4840]: I0930 15:00:25.070118 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-wmtv5_43ad4b75-0f40-4adc-8bda-67ed967051f3/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:00:25 crc kubenswrapper[4840]: I0930 15:00:25.148261 4840 generic.go:334] "Generic (PLEG): container finished" podID="518360cb-03c8-455f-8113-7c3a0493ff7c" containerID="94e5c91d166f6a6ab7e3247013f8e9bd52bad848b93f42b2c0817e2ad4ba9d2e" exitCode=0 Sep 30 15:00:25 crc kubenswrapper[4840]: I0930 15:00:25.148350 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4w8xk" event={"ID":"518360cb-03c8-455f-8113-7c3a0493ff7c","Type":"ContainerDied","Data":"94e5c91d166f6a6ab7e3247013f8e9bd52bad848b93f42b2c0817e2ad4ba9d2e"} Sep 30 15:00:26 crc kubenswrapper[4840]: I0930 15:00:26.161491 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4w8xk" event={"ID":"518360cb-03c8-455f-8113-7c3a0493ff7c","Type":"ContainerStarted","Data":"38eed2b6923624ad3c9371a6fc1b8605c385aef6f53207d978728977fb24a7a0"} Sep 30 15:00:26 crc kubenswrapper[4840]: I0930 15:00:26.189689 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4w8xk" podStartSLOduration=2.59383392 podStartE2EDuration="9.189665446s" podCreationTimestamp="2025-09-30 15:00:17 +0000 UTC" firstStartedPulling="2025-09-30 15:00:19.056899311 +0000 UTC m=+3847.685985744" lastFinishedPulling="2025-09-30 15:00:25.652730847 +0000 UTC m=+3854.281817270" observedRunningTime="2025-09-30 15:00:26.18277914 +0000 UTC m=+3854.811865573" watchObservedRunningTime="2025-09-30 15:00:26.189665446 +0000 UTC m=+3854.818751869" Sep 30 15:00:27 crc kubenswrapper[4840]: I0930 15:00:27.430795 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4w8xk" Sep 30 15:00:27 crc kubenswrapper[4840]: I0930 15:00:27.431202 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4w8xk" Sep 30 15:00:28 crc kubenswrapper[4840]: I0930 15:00:28.500905 4840 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4w8xk" podUID="518360cb-03c8-455f-8113-7c3a0493ff7c" containerName="registry-server" probeResult="failure" output=< Sep 30 15:00:28 crc kubenswrapper[4840]: timeout: failed to connect service ":50051" within 1s Sep 30 15:00:28 crc kubenswrapper[4840]: > Sep 30 15:00:35 crc kubenswrapper[4840]: I0930 15:00:35.375085 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_f0207213-00d5-4b63-b041-fff191463f4a/memcached/0.log" Sep 30 15:00:36 crc kubenswrapper[4840]: I0930 15:00:36.116301 4840 scope.go:117] "RemoveContainer" containerID="409430684b16cc7e5d86eb8dbc5c06065665c2c71a8402382eb805da2eca622c" Sep 30 15:00:36 crc kubenswrapper[4840]: E0930 15:00:36.116626 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 15:00:37 crc kubenswrapper[4840]: I0930 15:00:37.495497 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4w8xk" Sep 30 15:00:37 crc kubenswrapper[4840]: I0930 15:00:37.564032 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4w8xk" Sep 30 15:00:37 crc kubenswrapper[4840]: I0930 15:00:37.750300 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4w8xk"] Sep 30 15:00:39 crc kubenswrapper[4840]: I0930 15:00:39.285106 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4w8xk" podUID="518360cb-03c8-455f-8113-7c3a0493ff7c" containerName="registry-server" containerID="cri-o://38eed2b6923624ad3c9371a6fc1b8605c385aef6f53207d978728977fb24a7a0" gracePeriod=2 Sep 30 15:00:39 crc kubenswrapper[4840]: I0930 15:00:39.774052 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4w8xk" Sep 30 15:00:39 crc kubenswrapper[4840]: I0930 15:00:39.854787 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvr4r\" (UniqueName: \"kubernetes.io/projected/518360cb-03c8-455f-8113-7c3a0493ff7c-kube-api-access-qvr4r\") pod \"518360cb-03c8-455f-8113-7c3a0493ff7c\" (UID: \"518360cb-03c8-455f-8113-7c3a0493ff7c\") " Sep 30 15:00:39 crc kubenswrapper[4840]: I0930 15:00:39.854874 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/518360cb-03c8-455f-8113-7c3a0493ff7c-catalog-content\") pod \"518360cb-03c8-455f-8113-7c3a0493ff7c\" (UID: \"518360cb-03c8-455f-8113-7c3a0493ff7c\") " Sep 30 15:00:39 crc kubenswrapper[4840]: I0930 15:00:39.855006 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/518360cb-03c8-455f-8113-7c3a0493ff7c-utilities\") pod \"518360cb-03c8-455f-8113-7c3a0493ff7c\" (UID: \"518360cb-03c8-455f-8113-7c3a0493ff7c\") " Sep 30 15:00:39 crc kubenswrapper[4840]: I0930 15:00:39.855906 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/518360cb-03c8-455f-8113-7c3a0493ff7c-utilities" (OuterVolumeSpecName: "utilities") pod "518360cb-03c8-455f-8113-7c3a0493ff7c" (UID: "518360cb-03c8-455f-8113-7c3a0493ff7c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:00:39 crc kubenswrapper[4840]: I0930 15:00:39.863795 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/518360cb-03c8-455f-8113-7c3a0493ff7c-kube-api-access-qvr4r" (OuterVolumeSpecName: "kube-api-access-qvr4r") pod "518360cb-03c8-455f-8113-7c3a0493ff7c" (UID: "518360cb-03c8-455f-8113-7c3a0493ff7c"). InnerVolumeSpecName "kube-api-access-qvr4r". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:00:39 crc kubenswrapper[4840]: I0930 15:00:39.939489 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/518360cb-03c8-455f-8113-7c3a0493ff7c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "518360cb-03c8-455f-8113-7c3a0493ff7c" (UID: "518360cb-03c8-455f-8113-7c3a0493ff7c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:00:39 crc kubenswrapper[4840]: I0930 15:00:39.956768 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvr4r\" (UniqueName: \"kubernetes.io/projected/518360cb-03c8-455f-8113-7c3a0493ff7c-kube-api-access-qvr4r\") on node \"crc\" DevicePath \"\"" Sep 30 15:00:39 crc kubenswrapper[4840]: I0930 15:00:39.956816 4840 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/518360cb-03c8-455f-8113-7c3a0493ff7c-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 15:00:39 crc kubenswrapper[4840]: I0930 15:00:39.956828 4840 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/518360cb-03c8-455f-8113-7c3a0493ff7c-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 15:00:40 crc kubenswrapper[4840]: I0930 15:00:40.294361 4840 generic.go:334] "Generic (PLEG): container finished" podID="518360cb-03c8-455f-8113-7c3a0493ff7c" containerID="38eed2b6923624ad3c9371a6fc1b8605c385aef6f53207d978728977fb24a7a0" exitCode=0 Sep 30 15:00:40 crc kubenswrapper[4840]: I0930 15:00:40.294404 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4w8xk" Sep 30 15:00:40 crc kubenswrapper[4840]: I0930 15:00:40.294447 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4w8xk" event={"ID":"518360cb-03c8-455f-8113-7c3a0493ff7c","Type":"ContainerDied","Data":"38eed2b6923624ad3c9371a6fc1b8605c385aef6f53207d978728977fb24a7a0"} Sep 30 15:00:40 crc kubenswrapper[4840]: I0930 15:00:40.294793 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4w8xk" event={"ID":"518360cb-03c8-455f-8113-7c3a0493ff7c","Type":"ContainerDied","Data":"6cc60012f998341c5911c625e3aa767d3dc7e2e376aabc1e7ac2b45eeea78fe7"} Sep 30 15:00:40 crc kubenswrapper[4840]: I0930 15:00:40.294817 4840 scope.go:117] "RemoveContainer" containerID="38eed2b6923624ad3c9371a6fc1b8605c385aef6f53207d978728977fb24a7a0" Sep 30 15:00:40 crc kubenswrapper[4840]: I0930 15:00:40.312465 4840 scope.go:117] "RemoveContainer" containerID="94e5c91d166f6a6ab7e3247013f8e9bd52bad848b93f42b2c0817e2ad4ba9d2e" Sep 30 15:00:40 crc kubenswrapper[4840]: I0930 15:00:40.318169 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4w8xk"] Sep 30 15:00:40 crc kubenswrapper[4840]: I0930 15:00:40.328610 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4w8xk"] Sep 30 15:00:40 crc kubenswrapper[4840]: I0930 15:00:40.346962 4840 scope.go:117] "RemoveContainer" containerID="11e1516b31d5532eb60ffe13ceb9b893ebd6226cb5c0b72f908537b1f8e71ad3" Sep 30 15:00:40 crc kubenswrapper[4840]: I0930 15:00:40.387804 4840 scope.go:117] "RemoveContainer" containerID="38eed2b6923624ad3c9371a6fc1b8605c385aef6f53207d978728977fb24a7a0" Sep 30 15:00:40 crc kubenswrapper[4840]: E0930 15:00:40.388248 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38eed2b6923624ad3c9371a6fc1b8605c385aef6f53207d978728977fb24a7a0\": container with ID starting with 38eed2b6923624ad3c9371a6fc1b8605c385aef6f53207d978728977fb24a7a0 not found: ID does not exist" containerID="38eed2b6923624ad3c9371a6fc1b8605c385aef6f53207d978728977fb24a7a0" Sep 30 15:00:40 crc kubenswrapper[4840]: I0930 15:00:40.388285 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38eed2b6923624ad3c9371a6fc1b8605c385aef6f53207d978728977fb24a7a0"} err="failed to get container status \"38eed2b6923624ad3c9371a6fc1b8605c385aef6f53207d978728977fb24a7a0\": rpc error: code = NotFound desc = could not find container \"38eed2b6923624ad3c9371a6fc1b8605c385aef6f53207d978728977fb24a7a0\": container with ID starting with 38eed2b6923624ad3c9371a6fc1b8605c385aef6f53207d978728977fb24a7a0 not found: ID does not exist" Sep 30 15:00:40 crc kubenswrapper[4840]: I0930 15:00:40.388311 4840 scope.go:117] "RemoveContainer" containerID="94e5c91d166f6a6ab7e3247013f8e9bd52bad848b93f42b2c0817e2ad4ba9d2e" Sep 30 15:00:40 crc kubenswrapper[4840]: E0930 15:00:40.388710 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94e5c91d166f6a6ab7e3247013f8e9bd52bad848b93f42b2c0817e2ad4ba9d2e\": container with ID starting with 94e5c91d166f6a6ab7e3247013f8e9bd52bad848b93f42b2c0817e2ad4ba9d2e not found: ID does not exist" containerID="94e5c91d166f6a6ab7e3247013f8e9bd52bad848b93f42b2c0817e2ad4ba9d2e" Sep 30 15:00:40 crc kubenswrapper[4840]: I0930 15:00:40.388740 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94e5c91d166f6a6ab7e3247013f8e9bd52bad848b93f42b2c0817e2ad4ba9d2e"} err="failed to get container status \"94e5c91d166f6a6ab7e3247013f8e9bd52bad848b93f42b2c0817e2ad4ba9d2e\": rpc error: code = NotFound desc = could not find container \"94e5c91d166f6a6ab7e3247013f8e9bd52bad848b93f42b2c0817e2ad4ba9d2e\": container with ID starting with 94e5c91d166f6a6ab7e3247013f8e9bd52bad848b93f42b2c0817e2ad4ba9d2e not found: ID does not exist" Sep 30 15:00:40 crc kubenswrapper[4840]: I0930 15:00:40.388754 4840 scope.go:117] "RemoveContainer" containerID="11e1516b31d5532eb60ffe13ceb9b893ebd6226cb5c0b72f908537b1f8e71ad3" Sep 30 15:00:40 crc kubenswrapper[4840]: E0930 15:00:40.389250 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11e1516b31d5532eb60ffe13ceb9b893ebd6226cb5c0b72f908537b1f8e71ad3\": container with ID starting with 11e1516b31d5532eb60ffe13ceb9b893ebd6226cb5c0b72f908537b1f8e71ad3 not found: ID does not exist" containerID="11e1516b31d5532eb60ffe13ceb9b893ebd6226cb5c0b72f908537b1f8e71ad3" Sep 30 15:00:40 crc kubenswrapper[4840]: I0930 15:00:40.389294 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11e1516b31d5532eb60ffe13ceb9b893ebd6226cb5c0b72f908537b1f8e71ad3"} err="failed to get container status \"11e1516b31d5532eb60ffe13ceb9b893ebd6226cb5c0b72f908537b1f8e71ad3\": rpc error: code = NotFound desc = could not find container \"11e1516b31d5532eb60ffe13ceb9b893ebd6226cb5c0b72f908537b1f8e71ad3\": container with ID starting with 11e1516b31d5532eb60ffe13ceb9b893ebd6226cb5c0b72f908537b1f8e71ad3 not found: ID does not exist" Sep 30 15:00:42 crc kubenswrapper[4840]: I0930 15:00:42.127873 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="518360cb-03c8-455f-8113-7c3a0493ff7c" path="/var/lib/kubelet/pods/518360cb-03c8-455f-8113-7c3a0493ff7c/volumes" Sep 30 15:00:51 crc kubenswrapper[4840]: I0930 15:00:51.118348 4840 scope.go:117] "RemoveContainer" containerID="409430684b16cc7e5d86eb8dbc5c06065665c2c71a8402382eb805da2eca622c" Sep 30 15:00:51 crc kubenswrapper[4840]: E0930 15:00:51.119463 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 15:00:52 crc kubenswrapper[4840]: I0930 15:00:52.137638 4840 scope.go:117] "RemoveContainer" containerID="477513fbf915108bacd028dfd28bd853d6596414cf70f5430f7438625ae9e2e4" Sep 30 15:01:00 crc kubenswrapper[4840]: I0930 15:01:00.150939 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29320741-8tdtp"] Sep 30 15:01:00 crc kubenswrapper[4840]: E0930 15:01:00.151901 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="518360cb-03c8-455f-8113-7c3a0493ff7c" containerName="registry-server" Sep 30 15:01:00 crc kubenswrapper[4840]: I0930 15:01:00.151932 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="518360cb-03c8-455f-8113-7c3a0493ff7c" containerName="registry-server" Sep 30 15:01:00 crc kubenswrapper[4840]: E0930 15:01:00.151945 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="518360cb-03c8-455f-8113-7c3a0493ff7c" containerName="extract-utilities" Sep 30 15:01:00 crc kubenswrapper[4840]: I0930 15:01:00.151955 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="518360cb-03c8-455f-8113-7c3a0493ff7c" containerName="extract-utilities" Sep 30 15:01:00 crc kubenswrapper[4840]: E0930 15:01:00.151983 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="518360cb-03c8-455f-8113-7c3a0493ff7c" containerName="extract-content" Sep 30 15:01:00 crc kubenswrapper[4840]: I0930 15:01:00.151991 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="518360cb-03c8-455f-8113-7c3a0493ff7c" containerName="extract-content" Sep 30 15:01:00 crc kubenswrapper[4840]: I0930 15:01:00.152235 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="518360cb-03c8-455f-8113-7c3a0493ff7c" containerName="registry-server" Sep 30 15:01:00 crc kubenswrapper[4840]: I0930 15:01:00.153021 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29320741-8tdtp" Sep 30 15:01:00 crc kubenswrapper[4840]: I0930 15:01:00.164992 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29320741-8tdtp"] Sep 30 15:01:00 crc kubenswrapper[4840]: I0930 15:01:00.247822 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8857c244-058f-47f1-b865-85219c576636-config-data\") pod \"keystone-cron-29320741-8tdtp\" (UID: \"8857c244-058f-47f1-b865-85219c576636\") " pod="openstack/keystone-cron-29320741-8tdtp" Sep 30 15:01:00 crc kubenswrapper[4840]: I0930 15:01:00.248134 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wr4nr\" (UniqueName: \"kubernetes.io/projected/8857c244-058f-47f1-b865-85219c576636-kube-api-access-wr4nr\") pod \"keystone-cron-29320741-8tdtp\" (UID: \"8857c244-058f-47f1-b865-85219c576636\") " pod="openstack/keystone-cron-29320741-8tdtp" Sep 30 15:01:00 crc kubenswrapper[4840]: I0930 15:01:00.248293 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8857c244-058f-47f1-b865-85219c576636-fernet-keys\") pod \"keystone-cron-29320741-8tdtp\" (UID: \"8857c244-058f-47f1-b865-85219c576636\") " pod="openstack/keystone-cron-29320741-8tdtp" Sep 30 15:01:00 crc kubenswrapper[4840]: I0930 15:01:00.248343 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8857c244-058f-47f1-b865-85219c576636-combined-ca-bundle\") pod \"keystone-cron-29320741-8tdtp\" (UID: \"8857c244-058f-47f1-b865-85219c576636\") " pod="openstack/keystone-cron-29320741-8tdtp" Sep 30 15:01:00 crc kubenswrapper[4840]: I0930 15:01:00.349776 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8857c244-058f-47f1-b865-85219c576636-fernet-keys\") pod \"keystone-cron-29320741-8tdtp\" (UID: \"8857c244-058f-47f1-b865-85219c576636\") " pod="openstack/keystone-cron-29320741-8tdtp" Sep 30 15:01:00 crc kubenswrapper[4840]: I0930 15:01:00.349834 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8857c244-058f-47f1-b865-85219c576636-combined-ca-bundle\") pod \"keystone-cron-29320741-8tdtp\" (UID: \"8857c244-058f-47f1-b865-85219c576636\") " pod="openstack/keystone-cron-29320741-8tdtp" Sep 30 15:01:00 crc kubenswrapper[4840]: I0930 15:01:00.349927 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8857c244-058f-47f1-b865-85219c576636-config-data\") pod \"keystone-cron-29320741-8tdtp\" (UID: \"8857c244-058f-47f1-b865-85219c576636\") " pod="openstack/keystone-cron-29320741-8tdtp" Sep 30 15:01:00 crc kubenswrapper[4840]: I0930 15:01:00.349966 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wr4nr\" (UniqueName: \"kubernetes.io/projected/8857c244-058f-47f1-b865-85219c576636-kube-api-access-wr4nr\") pod \"keystone-cron-29320741-8tdtp\" (UID: \"8857c244-058f-47f1-b865-85219c576636\") " pod="openstack/keystone-cron-29320741-8tdtp" Sep 30 15:01:00 crc kubenswrapper[4840]: I0930 15:01:00.355482 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8857c244-058f-47f1-b865-85219c576636-combined-ca-bundle\") pod \"keystone-cron-29320741-8tdtp\" (UID: \"8857c244-058f-47f1-b865-85219c576636\") " pod="openstack/keystone-cron-29320741-8tdtp" Sep 30 15:01:00 crc kubenswrapper[4840]: I0930 15:01:00.357028 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8857c244-058f-47f1-b865-85219c576636-config-data\") pod \"keystone-cron-29320741-8tdtp\" (UID: \"8857c244-058f-47f1-b865-85219c576636\") " pod="openstack/keystone-cron-29320741-8tdtp" Sep 30 15:01:00 crc kubenswrapper[4840]: I0930 15:01:00.357587 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8857c244-058f-47f1-b865-85219c576636-fernet-keys\") pod \"keystone-cron-29320741-8tdtp\" (UID: \"8857c244-058f-47f1-b865-85219c576636\") " pod="openstack/keystone-cron-29320741-8tdtp" Sep 30 15:01:00 crc kubenswrapper[4840]: I0930 15:01:00.370868 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wr4nr\" (UniqueName: \"kubernetes.io/projected/8857c244-058f-47f1-b865-85219c576636-kube-api-access-wr4nr\") pod \"keystone-cron-29320741-8tdtp\" (UID: \"8857c244-058f-47f1-b865-85219c576636\") " pod="openstack/keystone-cron-29320741-8tdtp" Sep 30 15:01:00 crc kubenswrapper[4840]: I0930 15:01:00.476417 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29320741-8tdtp" Sep 30 15:01:00 crc kubenswrapper[4840]: I0930 15:01:00.518223 4840 generic.go:334] "Generic (PLEG): container finished" podID="7b5a37b7-2a02-4c6c-9d78-aa178cdf23c2" containerID="483b4c8eb835e622b0c96cb2dfbc37f655eb7ab76298701f6bc97c89ee6270f1" exitCode=0 Sep 30 15:01:00 crc kubenswrapper[4840]: I0930 15:01:00.518271 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gf4gj/crc-debug-x4l5h" event={"ID":"7b5a37b7-2a02-4c6c-9d78-aa178cdf23c2","Type":"ContainerDied","Data":"483b4c8eb835e622b0c96cb2dfbc37f655eb7ab76298701f6bc97c89ee6270f1"} Sep 30 15:01:00 crc kubenswrapper[4840]: I0930 15:01:00.917082 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29320741-8tdtp"] Sep 30 15:01:01 crc kubenswrapper[4840]: I0930 15:01:01.530487 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29320741-8tdtp" event={"ID":"8857c244-058f-47f1-b865-85219c576636","Type":"ContainerStarted","Data":"646fda1ef7f9777a80c14e6f459d572ef69188534ad2794e31f9b84c10352703"} Sep 30 15:01:01 crc kubenswrapper[4840]: I0930 15:01:01.530827 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29320741-8tdtp" event={"ID":"8857c244-058f-47f1-b865-85219c576636","Type":"ContainerStarted","Data":"85a04200aebb0be40ae6ffd09b1de612c5a712bcdd38b1157fcbdfc41119aea3"} Sep 30 15:01:01 crc kubenswrapper[4840]: I0930 15:01:01.548724 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29320741-8tdtp" podStartSLOduration=1.5487080340000001 podStartE2EDuration="1.548708034s" podCreationTimestamp="2025-09-30 15:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 15:01:01.544079936 +0000 UTC m=+3890.173166359" watchObservedRunningTime="2025-09-30 15:01:01.548708034 +0000 UTC m=+3890.177794457" Sep 30 15:01:01 crc kubenswrapper[4840]: I0930 15:01:01.641966 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gf4gj/crc-debug-x4l5h" Sep 30 15:01:01 crc kubenswrapper[4840]: I0930 15:01:01.671749 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-gf4gj/crc-debug-x4l5h"] Sep 30 15:01:01 crc kubenswrapper[4840]: I0930 15:01:01.680693 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-gf4gj/crc-debug-x4l5h"] Sep 30 15:01:01 crc kubenswrapper[4840]: I0930 15:01:01.778954 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njd95\" (UniqueName: \"kubernetes.io/projected/7b5a37b7-2a02-4c6c-9d78-aa178cdf23c2-kube-api-access-njd95\") pod \"7b5a37b7-2a02-4c6c-9d78-aa178cdf23c2\" (UID: \"7b5a37b7-2a02-4c6c-9d78-aa178cdf23c2\") " Sep 30 15:01:01 crc kubenswrapper[4840]: I0930 15:01:01.779166 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7b5a37b7-2a02-4c6c-9d78-aa178cdf23c2-host\") pod \"7b5a37b7-2a02-4c6c-9d78-aa178cdf23c2\" (UID: \"7b5a37b7-2a02-4c6c-9d78-aa178cdf23c2\") " Sep 30 15:01:01 crc kubenswrapper[4840]: I0930 15:01:01.780238 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b5a37b7-2a02-4c6c-9d78-aa178cdf23c2-host" (OuterVolumeSpecName: "host") pod "7b5a37b7-2a02-4c6c-9d78-aa178cdf23c2" (UID: "7b5a37b7-2a02-4c6c-9d78-aa178cdf23c2"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 15:01:01 crc kubenswrapper[4840]: I0930 15:01:01.792789 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b5a37b7-2a02-4c6c-9d78-aa178cdf23c2-kube-api-access-njd95" (OuterVolumeSpecName: "kube-api-access-njd95") pod "7b5a37b7-2a02-4c6c-9d78-aa178cdf23c2" (UID: "7b5a37b7-2a02-4c6c-9d78-aa178cdf23c2"). InnerVolumeSpecName "kube-api-access-njd95". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:01:01 crc kubenswrapper[4840]: I0930 15:01:01.881988 4840 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7b5a37b7-2a02-4c6c-9d78-aa178cdf23c2-host\") on node \"crc\" DevicePath \"\"" Sep 30 15:01:01 crc kubenswrapper[4840]: I0930 15:01:01.882031 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njd95\" (UniqueName: \"kubernetes.io/projected/7b5a37b7-2a02-4c6c-9d78-aa178cdf23c2-kube-api-access-njd95\") on node \"crc\" DevicePath \"\"" Sep 30 15:01:02 crc kubenswrapper[4840]: I0930 15:01:02.124222 4840 scope.go:117] "RemoveContainer" containerID="409430684b16cc7e5d86eb8dbc5c06065665c2c71a8402382eb805da2eca622c" Sep 30 15:01:02 crc kubenswrapper[4840]: E0930 15:01:02.124561 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 15:01:02 crc kubenswrapper[4840]: I0930 15:01:02.128678 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b5a37b7-2a02-4c6c-9d78-aa178cdf23c2" path="/var/lib/kubelet/pods/7b5a37b7-2a02-4c6c-9d78-aa178cdf23c2/volumes" Sep 30 15:01:02 crc kubenswrapper[4840]: I0930 15:01:02.540268 4840 scope.go:117] "RemoveContainer" containerID="483b4c8eb835e622b0c96cb2dfbc37f655eb7ab76298701f6bc97c89ee6270f1" Sep 30 15:01:02 crc kubenswrapper[4840]: I0930 15:01:02.540728 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gf4gj/crc-debug-x4l5h" Sep 30 15:01:02 crc kubenswrapper[4840]: I0930 15:01:02.860334 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-gf4gj/crc-debug-5js69"] Sep 30 15:01:02 crc kubenswrapper[4840]: E0930 15:01:02.861085 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b5a37b7-2a02-4c6c-9d78-aa178cdf23c2" containerName="container-00" Sep 30 15:01:02 crc kubenswrapper[4840]: I0930 15:01:02.861114 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b5a37b7-2a02-4c6c-9d78-aa178cdf23c2" containerName="container-00" Sep 30 15:01:02 crc kubenswrapper[4840]: I0930 15:01:02.861350 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b5a37b7-2a02-4c6c-9d78-aa178cdf23c2" containerName="container-00" Sep 30 15:01:02 crc kubenswrapper[4840]: I0930 15:01:02.862117 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gf4gj/crc-debug-5js69" Sep 30 15:01:03 crc kubenswrapper[4840]: I0930 15:01:03.001754 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkhhj\" (UniqueName: \"kubernetes.io/projected/603c1c87-5e38-4be3-b588-10e6f486311c-kube-api-access-jkhhj\") pod \"crc-debug-5js69\" (UID: \"603c1c87-5e38-4be3-b588-10e6f486311c\") " pod="openshift-must-gather-gf4gj/crc-debug-5js69" Sep 30 15:01:03 crc kubenswrapper[4840]: I0930 15:01:03.001823 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/603c1c87-5e38-4be3-b588-10e6f486311c-host\") pod \"crc-debug-5js69\" (UID: \"603c1c87-5e38-4be3-b588-10e6f486311c\") " pod="openshift-must-gather-gf4gj/crc-debug-5js69" Sep 30 15:01:03 crc kubenswrapper[4840]: I0930 15:01:03.103543 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkhhj\" (UniqueName: \"kubernetes.io/projected/603c1c87-5e38-4be3-b588-10e6f486311c-kube-api-access-jkhhj\") pod \"crc-debug-5js69\" (UID: \"603c1c87-5e38-4be3-b588-10e6f486311c\") " pod="openshift-must-gather-gf4gj/crc-debug-5js69" Sep 30 15:01:03 crc kubenswrapper[4840]: I0930 15:01:03.103660 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/603c1c87-5e38-4be3-b588-10e6f486311c-host\") pod \"crc-debug-5js69\" (UID: \"603c1c87-5e38-4be3-b588-10e6f486311c\") " pod="openshift-must-gather-gf4gj/crc-debug-5js69" Sep 30 15:01:03 crc kubenswrapper[4840]: I0930 15:01:03.103797 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/603c1c87-5e38-4be3-b588-10e6f486311c-host\") pod \"crc-debug-5js69\" (UID: \"603c1c87-5e38-4be3-b588-10e6f486311c\") " pod="openshift-must-gather-gf4gj/crc-debug-5js69" Sep 30 15:01:03 crc kubenswrapper[4840]: I0930 15:01:03.123521 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkhhj\" (UniqueName: \"kubernetes.io/projected/603c1c87-5e38-4be3-b588-10e6f486311c-kube-api-access-jkhhj\") pod \"crc-debug-5js69\" (UID: \"603c1c87-5e38-4be3-b588-10e6f486311c\") " pod="openshift-must-gather-gf4gj/crc-debug-5js69" Sep 30 15:01:03 crc kubenswrapper[4840]: I0930 15:01:03.180639 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gf4gj/crc-debug-5js69" Sep 30 15:01:03 crc kubenswrapper[4840]: W0930 15:01:03.209295 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod603c1c87_5e38_4be3_b588_10e6f486311c.slice/crio-b4dfe97aae40b50cae2c256f8558ef8010d1833eeb9ea658ed1a73bfb84e67e7 WatchSource:0}: Error finding container b4dfe97aae40b50cae2c256f8558ef8010d1833eeb9ea658ed1a73bfb84e67e7: Status 404 returned error can't find the container with id b4dfe97aae40b50cae2c256f8558ef8010d1833eeb9ea658ed1a73bfb84e67e7 Sep 30 15:01:03 crc kubenswrapper[4840]: I0930 15:01:03.550631 4840 generic.go:334] "Generic (PLEG): container finished" podID="8857c244-058f-47f1-b865-85219c576636" containerID="646fda1ef7f9777a80c14e6f459d572ef69188534ad2794e31f9b84c10352703" exitCode=0 Sep 30 15:01:03 crc kubenswrapper[4840]: I0930 15:01:03.550766 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29320741-8tdtp" event={"ID":"8857c244-058f-47f1-b865-85219c576636","Type":"ContainerDied","Data":"646fda1ef7f9777a80c14e6f459d572ef69188534ad2794e31f9b84c10352703"} Sep 30 15:01:03 crc kubenswrapper[4840]: I0930 15:01:03.552490 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gf4gj/crc-debug-5js69" event={"ID":"603c1c87-5e38-4be3-b588-10e6f486311c","Type":"ContainerStarted","Data":"741c591fbfe40904d124873f151df11afa1af159e42f0f2322aa6adf67e272d4"} Sep 30 15:01:03 crc kubenswrapper[4840]: I0930 15:01:03.552526 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gf4gj/crc-debug-5js69" event={"ID":"603c1c87-5e38-4be3-b588-10e6f486311c","Type":"ContainerStarted","Data":"b4dfe97aae40b50cae2c256f8558ef8010d1833eeb9ea658ed1a73bfb84e67e7"} Sep 30 15:01:03 crc kubenswrapper[4840]: I0930 15:01:03.585150 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-gf4gj/crc-debug-5js69" podStartSLOduration=1.585126082 podStartE2EDuration="1.585126082s" podCreationTimestamp="2025-09-30 15:01:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 15:01:03.580156485 +0000 UTC m=+3892.209242908" watchObservedRunningTime="2025-09-30 15:01:03.585126082 +0000 UTC m=+3892.214212505" Sep 30 15:01:04 crc kubenswrapper[4840]: I0930 15:01:04.568065 4840 generic.go:334] "Generic (PLEG): container finished" podID="603c1c87-5e38-4be3-b588-10e6f486311c" containerID="741c591fbfe40904d124873f151df11afa1af159e42f0f2322aa6adf67e272d4" exitCode=0 Sep 30 15:01:04 crc kubenswrapper[4840]: I0930 15:01:04.568122 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gf4gj/crc-debug-5js69" event={"ID":"603c1c87-5e38-4be3-b588-10e6f486311c","Type":"ContainerDied","Data":"741c591fbfe40904d124873f151df11afa1af159e42f0f2322aa6adf67e272d4"} Sep 30 15:01:04 crc kubenswrapper[4840]: I0930 15:01:04.900464 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29320741-8tdtp" Sep 30 15:01:05 crc kubenswrapper[4840]: I0930 15:01:05.030932 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wr4nr\" (UniqueName: \"kubernetes.io/projected/8857c244-058f-47f1-b865-85219c576636-kube-api-access-wr4nr\") pod \"8857c244-058f-47f1-b865-85219c576636\" (UID: \"8857c244-058f-47f1-b865-85219c576636\") " Sep 30 15:01:05 crc kubenswrapper[4840]: I0930 15:01:05.031006 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8857c244-058f-47f1-b865-85219c576636-combined-ca-bundle\") pod \"8857c244-058f-47f1-b865-85219c576636\" (UID: \"8857c244-058f-47f1-b865-85219c576636\") " Sep 30 15:01:05 crc kubenswrapper[4840]: I0930 15:01:05.031033 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8857c244-058f-47f1-b865-85219c576636-config-data\") pod \"8857c244-058f-47f1-b865-85219c576636\" (UID: \"8857c244-058f-47f1-b865-85219c576636\") " Sep 30 15:01:05 crc kubenswrapper[4840]: I0930 15:01:05.031054 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8857c244-058f-47f1-b865-85219c576636-fernet-keys\") pod \"8857c244-058f-47f1-b865-85219c576636\" (UID: \"8857c244-058f-47f1-b865-85219c576636\") " Sep 30 15:01:05 crc kubenswrapper[4840]: I0930 15:01:05.041894 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8857c244-058f-47f1-b865-85219c576636-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "8857c244-058f-47f1-b865-85219c576636" (UID: "8857c244-058f-47f1-b865-85219c576636"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 15:01:05 crc kubenswrapper[4840]: I0930 15:01:05.043809 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8857c244-058f-47f1-b865-85219c576636-kube-api-access-wr4nr" (OuterVolumeSpecName: "kube-api-access-wr4nr") pod "8857c244-058f-47f1-b865-85219c576636" (UID: "8857c244-058f-47f1-b865-85219c576636"). InnerVolumeSpecName "kube-api-access-wr4nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:01:05 crc kubenswrapper[4840]: I0930 15:01:05.082869 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8857c244-058f-47f1-b865-85219c576636-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8857c244-058f-47f1-b865-85219c576636" (UID: "8857c244-058f-47f1-b865-85219c576636"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 15:01:05 crc kubenswrapper[4840]: I0930 15:01:05.130666 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8857c244-058f-47f1-b865-85219c576636-config-data" (OuterVolumeSpecName: "config-data") pod "8857c244-058f-47f1-b865-85219c576636" (UID: "8857c244-058f-47f1-b865-85219c576636"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 15:01:05 crc kubenswrapper[4840]: I0930 15:01:05.145109 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wr4nr\" (UniqueName: \"kubernetes.io/projected/8857c244-058f-47f1-b865-85219c576636-kube-api-access-wr4nr\") on node \"crc\" DevicePath \"\"" Sep 30 15:01:05 crc kubenswrapper[4840]: I0930 15:01:05.145148 4840 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8857c244-058f-47f1-b865-85219c576636-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 15:01:05 crc kubenswrapper[4840]: I0930 15:01:05.145160 4840 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8857c244-058f-47f1-b865-85219c576636-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 15:01:05 crc kubenswrapper[4840]: I0930 15:01:05.145174 4840 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8857c244-058f-47f1-b865-85219c576636-fernet-keys\") on node \"crc\" DevicePath \"\"" Sep 30 15:01:05 crc kubenswrapper[4840]: I0930 15:01:05.583219 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29320741-8tdtp" Sep 30 15:01:05 crc kubenswrapper[4840]: I0930 15:01:05.583265 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29320741-8tdtp" event={"ID":"8857c244-058f-47f1-b865-85219c576636","Type":"ContainerDied","Data":"85a04200aebb0be40ae6ffd09b1de612c5a712bcdd38b1157fcbdfc41119aea3"} Sep 30 15:01:05 crc kubenswrapper[4840]: I0930 15:01:05.583306 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85a04200aebb0be40ae6ffd09b1de612c5a712bcdd38b1157fcbdfc41119aea3" Sep 30 15:01:05 crc kubenswrapper[4840]: I0930 15:01:05.602680 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gf4gj/crc-debug-5js69" Sep 30 15:01:05 crc kubenswrapper[4840]: I0930 15:01:05.658958 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/603c1c87-5e38-4be3-b588-10e6f486311c-host\") pod \"603c1c87-5e38-4be3-b588-10e6f486311c\" (UID: \"603c1c87-5e38-4be3-b588-10e6f486311c\") " Sep 30 15:01:05 crc kubenswrapper[4840]: I0930 15:01:05.659127 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkhhj\" (UniqueName: \"kubernetes.io/projected/603c1c87-5e38-4be3-b588-10e6f486311c-kube-api-access-jkhhj\") pod \"603c1c87-5e38-4be3-b588-10e6f486311c\" (UID: \"603c1c87-5e38-4be3-b588-10e6f486311c\") " Sep 30 15:01:05 crc kubenswrapper[4840]: I0930 15:01:05.659119 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/603c1c87-5e38-4be3-b588-10e6f486311c-host" (OuterVolumeSpecName: "host") pod "603c1c87-5e38-4be3-b588-10e6f486311c" (UID: "603c1c87-5e38-4be3-b588-10e6f486311c"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 15:01:05 crc kubenswrapper[4840]: I0930 15:01:05.660238 4840 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/603c1c87-5e38-4be3-b588-10e6f486311c-host\") on node \"crc\" DevicePath \"\"" Sep 30 15:01:05 crc kubenswrapper[4840]: I0930 15:01:05.663914 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/603c1c87-5e38-4be3-b588-10e6f486311c-kube-api-access-jkhhj" (OuterVolumeSpecName: "kube-api-access-jkhhj") pod "603c1c87-5e38-4be3-b588-10e6f486311c" (UID: "603c1c87-5e38-4be3-b588-10e6f486311c"). InnerVolumeSpecName "kube-api-access-jkhhj". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:01:05 crc kubenswrapper[4840]: I0930 15:01:05.764394 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkhhj\" (UniqueName: \"kubernetes.io/projected/603c1c87-5e38-4be3-b588-10e6f486311c-kube-api-access-jkhhj\") on node \"crc\" DevicePath \"\"" Sep 30 15:01:06 crc kubenswrapper[4840]: I0930 15:01:06.596362 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gf4gj/crc-debug-5js69" event={"ID":"603c1c87-5e38-4be3-b588-10e6f486311c","Type":"ContainerDied","Data":"b4dfe97aae40b50cae2c256f8558ef8010d1833eeb9ea658ed1a73bfb84e67e7"} Sep 30 15:01:06 crc kubenswrapper[4840]: I0930 15:01:06.596760 4840 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b4dfe97aae40b50cae2c256f8558ef8010d1833eeb9ea658ed1a73bfb84e67e7" Sep 30 15:01:06 crc kubenswrapper[4840]: I0930 15:01:06.596515 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gf4gj/crc-debug-5js69" Sep 30 15:01:10 crc kubenswrapper[4840]: I0930 15:01:10.163041 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-gf4gj/crc-debug-5js69"] Sep 30 15:01:10 crc kubenswrapper[4840]: I0930 15:01:10.163612 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-gf4gj/crc-debug-5js69"] Sep 30 15:01:11 crc kubenswrapper[4840]: I0930 15:01:11.373245 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-gf4gj/crc-debug-vmd4r"] Sep 30 15:01:11 crc kubenswrapper[4840]: E0930 15:01:11.373970 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="603c1c87-5e38-4be3-b588-10e6f486311c" containerName="container-00" Sep 30 15:01:11 crc kubenswrapper[4840]: I0930 15:01:11.373989 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="603c1c87-5e38-4be3-b588-10e6f486311c" containerName="container-00" Sep 30 15:01:11 crc kubenswrapper[4840]: E0930 15:01:11.374003 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8857c244-058f-47f1-b865-85219c576636" containerName="keystone-cron" Sep 30 15:01:11 crc kubenswrapper[4840]: I0930 15:01:11.374010 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="8857c244-058f-47f1-b865-85219c576636" containerName="keystone-cron" Sep 30 15:01:11 crc kubenswrapper[4840]: I0930 15:01:11.374242 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="603c1c87-5e38-4be3-b588-10e6f486311c" containerName="container-00" Sep 30 15:01:11 crc kubenswrapper[4840]: I0930 15:01:11.374280 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="8857c244-058f-47f1-b865-85219c576636" containerName="keystone-cron" Sep 30 15:01:11 crc kubenswrapper[4840]: I0930 15:01:11.375084 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gf4gj/crc-debug-vmd4r" Sep 30 15:01:11 crc kubenswrapper[4840]: I0930 15:01:11.460318 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cb9jx\" (UniqueName: \"kubernetes.io/projected/2da9d3fc-082b-4c16-8800-b7bde0545ce1-kube-api-access-cb9jx\") pod \"crc-debug-vmd4r\" (UID: \"2da9d3fc-082b-4c16-8800-b7bde0545ce1\") " pod="openshift-must-gather-gf4gj/crc-debug-vmd4r" Sep 30 15:01:11 crc kubenswrapper[4840]: I0930 15:01:11.460451 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2da9d3fc-082b-4c16-8800-b7bde0545ce1-host\") pod \"crc-debug-vmd4r\" (UID: \"2da9d3fc-082b-4c16-8800-b7bde0545ce1\") " pod="openshift-must-gather-gf4gj/crc-debug-vmd4r" Sep 30 15:01:11 crc kubenswrapper[4840]: I0930 15:01:11.562098 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cb9jx\" (UniqueName: \"kubernetes.io/projected/2da9d3fc-082b-4c16-8800-b7bde0545ce1-kube-api-access-cb9jx\") pod \"crc-debug-vmd4r\" (UID: \"2da9d3fc-082b-4c16-8800-b7bde0545ce1\") " pod="openshift-must-gather-gf4gj/crc-debug-vmd4r" Sep 30 15:01:11 crc kubenswrapper[4840]: I0930 15:01:11.562251 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2da9d3fc-082b-4c16-8800-b7bde0545ce1-host\") pod \"crc-debug-vmd4r\" (UID: \"2da9d3fc-082b-4c16-8800-b7bde0545ce1\") " pod="openshift-must-gather-gf4gj/crc-debug-vmd4r" Sep 30 15:01:11 crc kubenswrapper[4840]: I0930 15:01:11.562386 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2da9d3fc-082b-4c16-8800-b7bde0545ce1-host\") pod \"crc-debug-vmd4r\" (UID: \"2da9d3fc-082b-4c16-8800-b7bde0545ce1\") " pod="openshift-must-gather-gf4gj/crc-debug-vmd4r" Sep 30 15:01:11 crc kubenswrapper[4840]: I0930 15:01:11.580136 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cb9jx\" (UniqueName: \"kubernetes.io/projected/2da9d3fc-082b-4c16-8800-b7bde0545ce1-kube-api-access-cb9jx\") pod \"crc-debug-vmd4r\" (UID: \"2da9d3fc-082b-4c16-8800-b7bde0545ce1\") " pod="openshift-must-gather-gf4gj/crc-debug-vmd4r" Sep 30 15:01:11 crc kubenswrapper[4840]: I0930 15:01:11.693326 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gf4gj/crc-debug-vmd4r" Sep 30 15:01:12 crc kubenswrapper[4840]: I0930 15:01:12.131294 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="603c1c87-5e38-4be3-b588-10e6f486311c" path="/var/lib/kubelet/pods/603c1c87-5e38-4be3-b588-10e6f486311c/volumes" Sep 30 15:01:12 crc kubenswrapper[4840]: I0930 15:01:12.657470 4840 generic.go:334] "Generic (PLEG): container finished" podID="2da9d3fc-082b-4c16-8800-b7bde0545ce1" containerID="c317fbd37c948c77d73ae79b80c2ae33af5440fb17d6a86e6c920de36612e79e" exitCode=0 Sep 30 15:01:12 crc kubenswrapper[4840]: I0930 15:01:12.657610 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gf4gj/crc-debug-vmd4r" event={"ID":"2da9d3fc-082b-4c16-8800-b7bde0545ce1","Type":"ContainerDied","Data":"c317fbd37c948c77d73ae79b80c2ae33af5440fb17d6a86e6c920de36612e79e"} Sep 30 15:01:12 crc kubenswrapper[4840]: I0930 15:01:12.657852 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gf4gj/crc-debug-vmd4r" event={"ID":"2da9d3fc-082b-4c16-8800-b7bde0545ce1","Type":"ContainerStarted","Data":"2ba94f01c0cc7b897884591c2a95f4de89c5d9ad869f2f8199e0f3b0fe4a2dfb"} Sep 30 15:01:12 crc kubenswrapper[4840]: I0930 15:01:12.699670 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-gf4gj/crc-debug-vmd4r"] Sep 30 15:01:12 crc kubenswrapper[4840]: I0930 15:01:12.708261 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-gf4gj/crc-debug-vmd4r"] Sep 30 15:01:13 crc kubenswrapper[4840]: I0930 15:01:13.116491 4840 scope.go:117] "RemoveContainer" containerID="409430684b16cc7e5d86eb8dbc5c06065665c2c71a8402382eb805da2eca622c" Sep 30 15:01:13 crc kubenswrapper[4840]: E0930 15:01:13.116784 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 15:01:13 crc kubenswrapper[4840]: I0930 15:01:13.775533 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gf4gj/crc-debug-vmd4r" Sep 30 15:01:13 crc kubenswrapper[4840]: I0930 15:01:13.904764 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cb9jx\" (UniqueName: \"kubernetes.io/projected/2da9d3fc-082b-4c16-8800-b7bde0545ce1-kube-api-access-cb9jx\") pod \"2da9d3fc-082b-4c16-8800-b7bde0545ce1\" (UID: \"2da9d3fc-082b-4c16-8800-b7bde0545ce1\") " Sep 30 15:01:13 crc kubenswrapper[4840]: I0930 15:01:13.905183 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2da9d3fc-082b-4c16-8800-b7bde0545ce1-host\") pod \"2da9d3fc-082b-4c16-8800-b7bde0545ce1\" (UID: \"2da9d3fc-082b-4c16-8800-b7bde0545ce1\") " Sep 30 15:01:13 crc kubenswrapper[4840]: I0930 15:01:13.905303 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2da9d3fc-082b-4c16-8800-b7bde0545ce1-host" (OuterVolumeSpecName: "host") pod "2da9d3fc-082b-4c16-8800-b7bde0545ce1" (UID: "2da9d3fc-082b-4c16-8800-b7bde0545ce1"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 15:01:13 crc kubenswrapper[4840]: I0930 15:01:13.905817 4840 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2da9d3fc-082b-4c16-8800-b7bde0545ce1-host\") on node \"crc\" DevicePath \"\"" Sep 30 15:01:13 crc kubenswrapper[4840]: I0930 15:01:13.910803 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2da9d3fc-082b-4c16-8800-b7bde0545ce1-kube-api-access-cb9jx" (OuterVolumeSpecName: "kube-api-access-cb9jx") pod "2da9d3fc-082b-4c16-8800-b7bde0545ce1" (UID: "2da9d3fc-082b-4c16-8800-b7bde0545ce1"). InnerVolumeSpecName "kube-api-access-cb9jx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:01:14 crc kubenswrapper[4840]: I0930 15:01:14.008306 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cb9jx\" (UniqueName: \"kubernetes.io/projected/2da9d3fc-082b-4c16-8800-b7bde0545ce1-kube-api-access-cb9jx\") on node \"crc\" DevicePath \"\"" Sep 30 15:01:14 crc kubenswrapper[4840]: I0930 15:01:14.133919 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2da9d3fc-082b-4c16-8800-b7bde0545ce1" path="/var/lib/kubelet/pods/2da9d3fc-082b-4c16-8800-b7bde0545ce1/volumes" Sep 30 15:01:14 crc kubenswrapper[4840]: I0930 15:01:14.274973 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx_181ca13b-195f-45ea-9894-3383d192d377/util/0.log" Sep 30 15:01:14 crc kubenswrapper[4840]: I0930 15:01:14.435708 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx_181ca13b-195f-45ea-9894-3383d192d377/util/0.log" Sep 30 15:01:14 crc kubenswrapper[4840]: I0930 15:01:14.441194 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx_181ca13b-195f-45ea-9894-3383d192d377/pull/0.log" Sep 30 15:01:14 crc kubenswrapper[4840]: I0930 15:01:14.481992 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx_181ca13b-195f-45ea-9894-3383d192d377/pull/0.log" Sep 30 15:01:14 crc kubenswrapper[4840]: I0930 15:01:14.677251 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx_181ca13b-195f-45ea-9894-3383d192d377/pull/0.log" Sep 30 15:01:14 crc kubenswrapper[4840]: I0930 15:01:14.681754 4840 scope.go:117] "RemoveContainer" containerID="c317fbd37c948c77d73ae79b80c2ae33af5440fb17d6a86e6c920de36612e79e" Sep 30 15:01:14 crc kubenswrapper[4840]: I0930 15:01:14.682648 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gf4gj/crc-debug-vmd4r" Sep 30 15:01:14 crc kubenswrapper[4840]: I0930 15:01:14.683277 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx_181ca13b-195f-45ea-9894-3383d192d377/util/0.log" Sep 30 15:01:14 crc kubenswrapper[4840]: I0930 15:01:14.721879 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_06b7f53859b80301a450ea0b3f9c62adb0db51701088417995fe46a1695f8dx_181ca13b-195f-45ea-9894-3383d192d377/extract/0.log" Sep 30 15:01:14 crc kubenswrapper[4840]: I0930 15:01:14.871276 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-rnsbn_0f4e505b-4892-4c92-9ec4-0bdeb5d4c873/kube-rbac-proxy/0.log" Sep 30 15:01:14 crc kubenswrapper[4840]: I0930 15:01:14.922357 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-rnsbn_0f4e505b-4892-4c92-9ec4-0bdeb5d4c873/manager/0.log" Sep 30 15:01:14 crc kubenswrapper[4840]: I0930 15:01:14.967786 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-smg6p_3ffb0dd0-3dd8-4f37-8ead-4add746a1fc6/kube-rbac-proxy/0.log" Sep 30 15:01:15 crc kubenswrapper[4840]: I0930 15:01:15.285394 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-smg6p_3ffb0dd0-3dd8-4f37-8ead-4add746a1fc6/manager/0.log" Sep 30 15:01:15 crc kubenswrapper[4840]: I0930 15:01:15.308394 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-6nn29_395025e2-9d12-4172-bb73-b4fb67e0111c/kube-rbac-proxy/0.log" Sep 30 15:01:15 crc kubenswrapper[4840]: I0930 15:01:15.371889 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-6nn29_395025e2-9d12-4172-bb73-b4fb67e0111c/manager/0.log" Sep 30 15:01:15 crc kubenswrapper[4840]: I0930 15:01:15.489841 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-sq5hd_64c5ad91-f113-46e3-ac87-36521d32f4f0/kube-rbac-proxy/0.log" Sep 30 15:01:15 crc kubenswrapper[4840]: I0930 15:01:15.581163 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-sq5hd_64c5ad91-f113-46e3-ac87-36521d32f4f0/manager/0.log" Sep 30 15:01:15 crc kubenswrapper[4840]: I0930 15:01:15.696210 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-jwlj9_5daf5662-8848-4fd9-a6af-e091fe303d89/kube-rbac-proxy/0.log" Sep 30 15:01:15 crc kubenswrapper[4840]: I0930 15:01:15.741080 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-jwlj9_5daf5662-8848-4fd9-a6af-e091fe303d89/manager/0.log" Sep 30 15:01:15 crc kubenswrapper[4840]: I0930 15:01:15.779416 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-d9jgv_63d94c36-0a55-472c-bbe8-27195105584e/kube-rbac-proxy/0.log" Sep 30 15:01:15 crc kubenswrapper[4840]: I0930 15:01:15.899811 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-d9jgv_63d94c36-0a55-472c-bbe8-27195105584e/manager/0.log" Sep 30 15:01:15 crc kubenswrapper[4840]: I0930 15:01:15.973028 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-7d857cc749-xtqwd_807f2091-02fa-41e0-943a-b30f4202c476/kube-rbac-proxy/0.log" Sep 30 15:01:16 crc kubenswrapper[4840]: I0930 15:01:16.165699 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-7d857cc749-xtqwd_807f2091-02fa-41e0-943a-b30f4202c476/manager/0.log" Sep 30 15:01:16 crc kubenswrapper[4840]: I0930 15:01:16.200016 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-7975b88857-fg7c7_494a382c-6fec-4711-8a90-7f691c847e86/kube-rbac-proxy/0.log" Sep 30 15:01:16 crc kubenswrapper[4840]: I0930 15:01:16.211403 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-7975b88857-fg7c7_494a382c-6fec-4711-8a90-7f691c847e86/manager/0.log" Sep 30 15:01:16 crc kubenswrapper[4840]: I0930 15:01:16.368207 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5bd55b4bff-rjpbm_b927784b-0b44-4bf1-9751-3aafcc24367b/kube-rbac-proxy/0.log" Sep 30 15:01:16 crc kubenswrapper[4840]: I0930 15:01:16.432380 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5bd55b4bff-rjpbm_b927784b-0b44-4bf1-9751-3aafcc24367b/manager/0.log" Sep 30 15:01:16 crc kubenswrapper[4840]: I0930 15:01:16.581109 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-d2hgd_eb939f27-ce94-408d-b673-4c25f96dd059/manager/0.log" Sep 30 15:01:16 crc kubenswrapper[4840]: I0930 15:01:16.588177 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-d2hgd_eb939f27-ce94-408d-b673-4c25f96dd059/kube-rbac-proxy/0.log" Sep 30 15:01:16 crc kubenswrapper[4840]: I0930 15:01:16.686193 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-xzmg8_385365f0-9e2c-4c89-b95b-164a473ff0c3/kube-rbac-proxy/0.log" Sep 30 15:01:16 crc kubenswrapper[4840]: I0930 15:01:16.776659 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-xzmg8_385365f0-9e2c-4c89-b95b-164a473ff0c3/manager/0.log" Sep 30 15:01:16 crc kubenswrapper[4840]: I0930 15:01:16.887917 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-64d7b59854-qsq84_3d5da746-e5f3-4bed-9c77-37a78da8f1a6/kube-rbac-proxy/0.log" Sep 30 15:01:16 crc kubenswrapper[4840]: I0930 15:01:16.987780 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-64d7b59854-qsq84_3d5da746-e5f3-4bed-9c77-37a78da8f1a6/manager/0.log" Sep 30 15:01:17 crc kubenswrapper[4840]: I0930 15:01:17.016548 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-c7c776c96-x4mx4_3f4d112b-e6b8-455a-9f66-e8ee1bb81ac9/kube-rbac-proxy/0.log" Sep 30 15:01:17 crc kubenswrapper[4840]: I0930 15:01:17.164451 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-c7c776c96-x4mx4_3f4d112b-e6b8-455a-9f66-e8ee1bb81ac9/manager/0.log" Sep 30 15:01:17 crc kubenswrapper[4840]: I0930 15:01:17.217767 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-76fcc6dc7c-t2mx4_a76ab4ec-b180-473f-a597-641725526b58/kube-rbac-proxy/0.log" Sep 30 15:01:17 crc kubenswrapper[4840]: I0930 15:01:17.234653 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-76fcc6dc7c-t2mx4_a76ab4ec-b180-473f-a597-641725526b58/manager/0.log" Sep 30 15:01:17 crc kubenswrapper[4840]: I0930 15:01:17.373919 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6d776955-p6msl_62475ec3-b883-4d16-bb31-6ee393062da2/kube-rbac-proxy/0.log" Sep 30 15:01:17 crc kubenswrapper[4840]: I0930 15:01:17.391030 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6d776955-p6msl_62475ec3-b883-4d16-bb31-6ee393062da2/manager/0.log" Sep 30 15:01:17 crc kubenswrapper[4840]: I0930 15:01:17.567776 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-568796b78d-bvbdn_e61dcaab-7705-4857-bdd9-adeb07d9e6f2/kube-rbac-proxy/0.log" Sep 30 15:01:17 crc kubenswrapper[4840]: I0930 15:01:17.691425 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-84594b7f86-xwqzh_0b30b057-e591-411e-919b-5d1efac1a249/kube-rbac-proxy/0.log" Sep 30 15:01:17 crc kubenswrapper[4840]: I0930 15:01:17.902754 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-84594b7f86-xwqzh_0b30b057-e591-411e-919b-5d1efac1a249/operator/0.log" Sep 30 15:01:17 crc kubenswrapper[4840]: I0930 15:01:17.905431 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-gl8hl_a1690286-e0ea-4cf7-877b-d5e172dc365e/registry-server/0.log" Sep 30 15:01:18 crc kubenswrapper[4840]: I0930 15:01:18.073996 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-9t25m_dc75c0e8-691e-4108-b9c7-e4233d0baeb6/kube-rbac-proxy/0.log" Sep 30 15:01:18 crc kubenswrapper[4840]: I0930 15:01:18.181742 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-9t25m_dc75c0e8-691e-4108-b9c7-e4233d0baeb6/manager/0.log" Sep 30 15:01:18 crc kubenswrapper[4840]: I0930 15:01:18.233680 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-tqjvl_3ed332eb-b604-4667-90f4-541d5a9115c2/kube-rbac-proxy/0.log" Sep 30 15:01:18 crc kubenswrapper[4840]: I0930 15:01:18.450316 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-tqjvl_3ed332eb-b604-4667-90f4-541d5a9115c2/manager/0.log" Sep 30 15:01:18 crc kubenswrapper[4840]: I0930 15:01:18.498353 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-79d8469568-smjg7_aa2cf241-ec41-4a94-b452-653b3439a073/operator/0.log" Sep 30 15:01:18 crc kubenswrapper[4840]: I0930 15:01:18.666329 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-bc7dc7bd9-4mpb8_2b8703c9-f691-4c3a-9941-8bcd36aaf948/kube-rbac-proxy/0.log" Sep 30 15:01:18 crc kubenswrapper[4840]: I0930 15:01:18.729087 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-bc7dc7bd9-4mpb8_2b8703c9-f691-4c3a-9941-8bcd36aaf948/manager/0.log" Sep 30 15:01:18 crc kubenswrapper[4840]: I0930 15:01:18.752264 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-568796b78d-bvbdn_e61dcaab-7705-4857-bdd9-adeb07d9e6f2/manager/0.log" Sep 30 15:01:18 crc kubenswrapper[4840]: I0930 15:01:18.847331 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-78974455b9-97rs4_47248483-6556-491b-b65f-01cad9f1d1fc/kube-rbac-proxy/0.log" Sep 30 15:01:18 crc kubenswrapper[4840]: I0930 15:01:18.933889 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-78974455b9-97rs4_47248483-6556-491b-b65f-01cad9f1d1fc/manager/0.log" Sep 30 15:01:18 crc kubenswrapper[4840]: I0930 15:01:18.984749 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-f66b554c6-9npm5_3ba6da78-9606-4e1c-bc55-6cbe2e052464/kube-rbac-proxy/0.log" Sep 30 15:01:18 crc kubenswrapper[4840]: I0930 15:01:18.990354 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-f66b554c6-9npm5_3ba6da78-9606-4e1c-bc55-6cbe2e052464/manager/0.log" Sep 30 15:01:19 crc kubenswrapper[4840]: I0930 15:01:19.119700 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-76669f99c-rn2qq_e1cd519e-b5c3-48f2-b426-d0f4e912683b/kube-rbac-proxy/0.log" Sep 30 15:01:19 crc kubenswrapper[4840]: I0930 15:01:19.148945 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-76669f99c-rn2qq_e1cd519e-b5c3-48f2-b426-d0f4e912683b/manager/0.log" Sep 30 15:01:28 crc kubenswrapper[4840]: I0930 15:01:28.117117 4840 scope.go:117] "RemoveContainer" containerID="409430684b16cc7e5d86eb8dbc5c06065665c2c71a8402382eb805da2eca622c" Sep 30 15:01:28 crc kubenswrapper[4840]: E0930 15:01:28.117973 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 15:01:33 crc kubenswrapper[4840]: I0930 15:01:33.692437 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-z866q_eb59a261-a39c-4be5-86f0-a14643aef960/control-plane-machine-set-operator/0.log" Sep 30 15:01:33 crc kubenswrapper[4840]: I0930 15:01:33.885235 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-zbnls_d3508091-c5c6-4dca-9b32-883426154a6e/kube-rbac-proxy/0.log" Sep 30 15:01:33 crc kubenswrapper[4840]: I0930 15:01:33.916233 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-zbnls_d3508091-c5c6-4dca-9b32-883426154a6e/machine-api-operator/0.log" Sep 30 15:01:38 crc kubenswrapper[4840]: I0930 15:01:38.810552 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-78zcs"] Sep 30 15:01:38 crc kubenswrapper[4840]: E0930 15:01:38.811543 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2da9d3fc-082b-4c16-8800-b7bde0545ce1" containerName="container-00" Sep 30 15:01:38 crc kubenswrapper[4840]: I0930 15:01:38.811567 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="2da9d3fc-082b-4c16-8800-b7bde0545ce1" containerName="container-00" Sep 30 15:01:38 crc kubenswrapper[4840]: I0930 15:01:38.811837 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="2da9d3fc-082b-4c16-8800-b7bde0545ce1" containerName="container-00" Sep 30 15:01:38 crc kubenswrapper[4840]: I0930 15:01:38.813272 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-78zcs" Sep 30 15:01:38 crc kubenswrapper[4840]: I0930 15:01:38.819674 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-78zcs"] Sep 30 15:01:38 crc kubenswrapper[4840]: I0930 15:01:38.966646 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjzzs\" (UniqueName: \"kubernetes.io/projected/9a7d4b45-82ea-4041-8391-f96864afea7a-kube-api-access-pjzzs\") pod \"certified-operators-78zcs\" (UID: \"9a7d4b45-82ea-4041-8391-f96864afea7a\") " pod="openshift-marketplace/certified-operators-78zcs" Sep 30 15:01:38 crc kubenswrapper[4840]: I0930 15:01:38.966708 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a7d4b45-82ea-4041-8391-f96864afea7a-catalog-content\") pod \"certified-operators-78zcs\" (UID: \"9a7d4b45-82ea-4041-8391-f96864afea7a\") " pod="openshift-marketplace/certified-operators-78zcs" Sep 30 15:01:38 crc kubenswrapper[4840]: I0930 15:01:38.966733 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a7d4b45-82ea-4041-8391-f96864afea7a-utilities\") pod \"certified-operators-78zcs\" (UID: \"9a7d4b45-82ea-4041-8391-f96864afea7a\") " pod="openshift-marketplace/certified-operators-78zcs" Sep 30 15:01:39 crc kubenswrapper[4840]: I0930 15:01:39.069066 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjzzs\" (UniqueName: \"kubernetes.io/projected/9a7d4b45-82ea-4041-8391-f96864afea7a-kube-api-access-pjzzs\") pod \"certified-operators-78zcs\" (UID: \"9a7d4b45-82ea-4041-8391-f96864afea7a\") " pod="openshift-marketplace/certified-operators-78zcs" Sep 30 15:01:39 crc kubenswrapper[4840]: I0930 15:01:39.069152 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a7d4b45-82ea-4041-8391-f96864afea7a-catalog-content\") pod \"certified-operators-78zcs\" (UID: \"9a7d4b45-82ea-4041-8391-f96864afea7a\") " pod="openshift-marketplace/certified-operators-78zcs" Sep 30 15:01:39 crc kubenswrapper[4840]: I0930 15:01:39.069177 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a7d4b45-82ea-4041-8391-f96864afea7a-utilities\") pod \"certified-operators-78zcs\" (UID: \"9a7d4b45-82ea-4041-8391-f96864afea7a\") " pod="openshift-marketplace/certified-operators-78zcs" Sep 30 15:01:39 crc kubenswrapper[4840]: I0930 15:01:39.069998 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a7d4b45-82ea-4041-8391-f96864afea7a-catalog-content\") pod \"certified-operators-78zcs\" (UID: \"9a7d4b45-82ea-4041-8391-f96864afea7a\") " pod="openshift-marketplace/certified-operators-78zcs" Sep 30 15:01:39 crc kubenswrapper[4840]: I0930 15:01:39.070029 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a7d4b45-82ea-4041-8391-f96864afea7a-utilities\") pod \"certified-operators-78zcs\" (UID: \"9a7d4b45-82ea-4041-8391-f96864afea7a\") " pod="openshift-marketplace/certified-operators-78zcs" Sep 30 15:01:39 crc kubenswrapper[4840]: I0930 15:01:39.094487 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjzzs\" (UniqueName: \"kubernetes.io/projected/9a7d4b45-82ea-4041-8391-f96864afea7a-kube-api-access-pjzzs\") pod \"certified-operators-78zcs\" (UID: \"9a7d4b45-82ea-4041-8391-f96864afea7a\") " pod="openshift-marketplace/certified-operators-78zcs" Sep 30 15:01:39 crc kubenswrapper[4840]: I0930 15:01:39.149277 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-78zcs" Sep 30 15:01:39 crc kubenswrapper[4840]: I0930 15:01:39.640320 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-78zcs"] Sep 30 15:01:39 crc kubenswrapper[4840]: W0930 15:01:39.644139 4840 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9a7d4b45_82ea_4041_8391_f96864afea7a.slice/crio-f3090a11940efd6e2891856c045be039de007ea8d6453f60401ea85bda1aba06 WatchSource:0}: Error finding container f3090a11940efd6e2891856c045be039de007ea8d6453f60401ea85bda1aba06: Status 404 returned error can't find the container with id f3090a11940efd6e2891856c045be039de007ea8d6453f60401ea85bda1aba06 Sep 30 15:01:39 crc kubenswrapper[4840]: I0930 15:01:39.905954 4840 generic.go:334] "Generic (PLEG): container finished" podID="9a7d4b45-82ea-4041-8391-f96864afea7a" containerID="3be2f6fe08d489ce0badbc260b6ad3e847057e9c612445dafad1f5b7f82dfe5c" exitCode=0 Sep 30 15:01:39 crc kubenswrapper[4840]: I0930 15:01:39.906032 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-78zcs" event={"ID":"9a7d4b45-82ea-4041-8391-f96864afea7a","Type":"ContainerDied","Data":"3be2f6fe08d489ce0badbc260b6ad3e847057e9c612445dafad1f5b7f82dfe5c"} Sep 30 15:01:39 crc kubenswrapper[4840]: I0930 15:01:39.906270 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-78zcs" event={"ID":"9a7d4b45-82ea-4041-8391-f96864afea7a","Type":"ContainerStarted","Data":"f3090a11940efd6e2891856c045be039de007ea8d6453f60401ea85bda1aba06"} Sep 30 15:01:41 crc kubenswrapper[4840]: I0930 15:01:41.939938 4840 generic.go:334] "Generic (PLEG): container finished" podID="9a7d4b45-82ea-4041-8391-f96864afea7a" containerID="2d7c7d7f24a7d676ceae64b2d7ce10d2f50befd9de34bc13ce9a2e291405b3ec" exitCode=0 Sep 30 15:01:41 crc kubenswrapper[4840]: I0930 15:01:41.940628 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-78zcs" event={"ID":"9a7d4b45-82ea-4041-8391-f96864afea7a","Type":"ContainerDied","Data":"2d7c7d7f24a7d676ceae64b2d7ce10d2f50befd9de34bc13ce9a2e291405b3ec"} Sep 30 15:01:43 crc kubenswrapper[4840]: I0930 15:01:43.116439 4840 scope.go:117] "RemoveContainer" containerID="409430684b16cc7e5d86eb8dbc5c06065665c2c71a8402382eb805da2eca622c" Sep 30 15:01:43 crc kubenswrapper[4840]: E0930 15:01:43.116828 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 15:01:43 crc kubenswrapper[4840]: I0930 15:01:43.960380 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-78zcs" event={"ID":"9a7d4b45-82ea-4041-8391-f96864afea7a","Type":"ContainerStarted","Data":"170bad059641917deb425e256a163286320331fadbebc1820ae78cd7cbd14a72"} Sep 30 15:01:43 crc kubenswrapper[4840]: I0930 15:01:43.986298 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-78zcs" podStartSLOduration=2.453405308 podStartE2EDuration="5.986280509s" podCreationTimestamp="2025-09-30 15:01:38 +0000 UTC" firstStartedPulling="2025-09-30 15:01:39.908742471 +0000 UTC m=+3928.537828894" lastFinishedPulling="2025-09-30 15:01:43.441617672 +0000 UTC m=+3932.070704095" observedRunningTime="2025-09-30 15:01:43.984735249 +0000 UTC m=+3932.613821682" watchObservedRunningTime="2025-09-30 15:01:43.986280509 +0000 UTC m=+3932.615366932" Sep 30 15:01:45 crc kubenswrapper[4840]: I0930 15:01:45.135653 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-hjfs2_11b92f8e-b36f-47e5-8d1a-d3c83ccbd870/cert-manager-controller/0.log" Sep 30 15:01:45 crc kubenswrapper[4840]: I0930 15:01:45.335812 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-f4kpb_8df1331d-54ff-47cd-80c9-c18a5ab2df53/cert-manager-cainjector/0.log" Sep 30 15:01:45 crc kubenswrapper[4840]: I0930 15:01:45.354194 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-sszwr_18d21186-8397-400e-a9df-5e4e41e6739e/cert-manager-webhook/0.log" Sep 30 15:01:49 crc kubenswrapper[4840]: I0930 15:01:49.150259 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-78zcs" Sep 30 15:01:49 crc kubenswrapper[4840]: I0930 15:01:49.150512 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-78zcs" Sep 30 15:01:49 crc kubenswrapper[4840]: I0930 15:01:49.329968 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-78zcs" Sep 30 15:01:50 crc kubenswrapper[4840]: I0930 15:01:50.070277 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-78zcs" Sep 30 15:01:50 crc kubenswrapper[4840]: I0930 15:01:50.132240 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-78zcs"] Sep 30 15:01:52 crc kubenswrapper[4840]: I0930 15:01:52.040089 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-78zcs" podUID="9a7d4b45-82ea-4041-8391-f96864afea7a" containerName="registry-server" containerID="cri-o://170bad059641917deb425e256a163286320331fadbebc1820ae78cd7cbd14a72" gracePeriod=2 Sep 30 15:01:52 crc kubenswrapper[4840]: I0930 15:01:52.204485 4840 scope.go:117] "RemoveContainer" containerID="aa4a55fc94ddbc1c23a4fb96abd1fbde1aa569ecad5702aabc268334a51cdc8b" Sep 30 15:01:53 crc kubenswrapper[4840]: I0930 15:01:53.062819 4840 generic.go:334] "Generic (PLEG): container finished" podID="9a7d4b45-82ea-4041-8391-f96864afea7a" containerID="170bad059641917deb425e256a163286320331fadbebc1820ae78cd7cbd14a72" exitCode=0 Sep 30 15:01:53 crc kubenswrapper[4840]: I0930 15:01:53.062871 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-78zcs" event={"ID":"9a7d4b45-82ea-4041-8391-f96864afea7a","Type":"ContainerDied","Data":"170bad059641917deb425e256a163286320331fadbebc1820ae78cd7cbd14a72"} Sep 30 15:01:53 crc kubenswrapper[4840]: I0930 15:01:53.278734 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-78zcs" Sep 30 15:01:53 crc kubenswrapper[4840]: I0930 15:01:53.347873 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjzzs\" (UniqueName: \"kubernetes.io/projected/9a7d4b45-82ea-4041-8391-f96864afea7a-kube-api-access-pjzzs\") pod \"9a7d4b45-82ea-4041-8391-f96864afea7a\" (UID: \"9a7d4b45-82ea-4041-8391-f96864afea7a\") " Sep 30 15:01:53 crc kubenswrapper[4840]: I0930 15:01:53.347947 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a7d4b45-82ea-4041-8391-f96864afea7a-catalog-content\") pod \"9a7d4b45-82ea-4041-8391-f96864afea7a\" (UID: \"9a7d4b45-82ea-4041-8391-f96864afea7a\") " Sep 30 15:01:53 crc kubenswrapper[4840]: I0930 15:01:53.347984 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a7d4b45-82ea-4041-8391-f96864afea7a-utilities\") pod \"9a7d4b45-82ea-4041-8391-f96864afea7a\" (UID: \"9a7d4b45-82ea-4041-8391-f96864afea7a\") " Sep 30 15:01:53 crc kubenswrapper[4840]: I0930 15:01:53.349038 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a7d4b45-82ea-4041-8391-f96864afea7a-utilities" (OuterVolumeSpecName: "utilities") pod "9a7d4b45-82ea-4041-8391-f96864afea7a" (UID: "9a7d4b45-82ea-4041-8391-f96864afea7a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:01:53 crc kubenswrapper[4840]: I0930 15:01:53.352903 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a7d4b45-82ea-4041-8391-f96864afea7a-kube-api-access-pjzzs" (OuterVolumeSpecName: "kube-api-access-pjzzs") pod "9a7d4b45-82ea-4041-8391-f96864afea7a" (UID: "9a7d4b45-82ea-4041-8391-f96864afea7a"). InnerVolumeSpecName "kube-api-access-pjzzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:01:53 crc kubenswrapper[4840]: I0930 15:01:53.395101 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a7d4b45-82ea-4041-8391-f96864afea7a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9a7d4b45-82ea-4041-8391-f96864afea7a" (UID: "9a7d4b45-82ea-4041-8391-f96864afea7a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:01:53 crc kubenswrapper[4840]: I0930 15:01:53.451246 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjzzs\" (UniqueName: \"kubernetes.io/projected/9a7d4b45-82ea-4041-8391-f96864afea7a-kube-api-access-pjzzs\") on node \"crc\" DevicePath \"\"" Sep 30 15:01:53 crc kubenswrapper[4840]: I0930 15:01:53.451299 4840 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a7d4b45-82ea-4041-8391-f96864afea7a-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 15:01:53 crc kubenswrapper[4840]: I0930 15:01:53.451311 4840 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a7d4b45-82ea-4041-8391-f96864afea7a-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 15:01:54 crc kubenswrapper[4840]: I0930 15:01:54.071771 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-78zcs" event={"ID":"9a7d4b45-82ea-4041-8391-f96864afea7a","Type":"ContainerDied","Data":"f3090a11940efd6e2891856c045be039de007ea8d6453f60401ea85bda1aba06"} Sep 30 15:01:54 crc kubenswrapper[4840]: I0930 15:01:54.071827 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-78zcs" Sep 30 15:01:54 crc kubenswrapper[4840]: I0930 15:01:54.071839 4840 scope.go:117] "RemoveContainer" containerID="170bad059641917deb425e256a163286320331fadbebc1820ae78cd7cbd14a72" Sep 30 15:01:54 crc kubenswrapper[4840]: I0930 15:01:54.100416 4840 scope.go:117] "RemoveContainer" containerID="2d7c7d7f24a7d676ceae64b2d7ce10d2f50befd9de34bc13ce9a2e291405b3ec" Sep 30 15:01:54 crc kubenswrapper[4840]: I0930 15:01:54.108013 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-78zcs"] Sep 30 15:01:54 crc kubenswrapper[4840]: I0930 15:01:54.128195 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-78zcs"] Sep 30 15:01:54 crc kubenswrapper[4840]: I0930 15:01:54.138439 4840 scope.go:117] "RemoveContainer" containerID="3be2f6fe08d489ce0badbc260b6ad3e847057e9c612445dafad1f5b7f82dfe5c" Sep 30 15:01:55 crc kubenswrapper[4840]: I0930 15:01:55.117121 4840 scope.go:117] "RemoveContainer" containerID="409430684b16cc7e5d86eb8dbc5c06065665c2c71a8402382eb805da2eca622c" Sep 30 15:01:55 crc kubenswrapper[4840]: E0930 15:01:55.117450 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 15:01:56 crc kubenswrapper[4840]: I0930 15:01:56.130432 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a7d4b45-82ea-4041-8391-f96864afea7a" path="/var/lib/kubelet/pods/9a7d4b45-82ea-4041-8391-f96864afea7a/volumes" Sep 30 15:01:57 crc kubenswrapper[4840]: I0930 15:01:57.511915 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-864bb6dfb5-lw82x_14ccc5bb-b26a-42df-b0b9-f1ea9bcbff31/nmstate-console-plugin/0.log" Sep 30 15:01:57 crc kubenswrapper[4840]: I0930 15:01:57.699892 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-gnz9p_7e3e89d5-05d9-4675-9dad-83f59258f876/nmstate-handler/0.log" Sep 30 15:01:57 crc kubenswrapper[4840]: I0930 15:01:57.807133 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58fcddf996-bw56r_9ed06022-f9fe-4b61-bd42-ed308118a717/kube-rbac-proxy/0.log" Sep 30 15:01:57 crc kubenswrapper[4840]: I0930 15:01:57.839341 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58fcddf996-bw56r_9ed06022-f9fe-4b61-bd42-ed308118a717/nmstate-metrics/0.log" Sep 30 15:01:57 crc kubenswrapper[4840]: I0930 15:01:57.967567 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5d6f6cfd66-wqtsv_5665a62e-e1b7-4773-afcb-009f7ffb354d/nmstate-operator/0.log" Sep 30 15:01:58 crc kubenswrapper[4840]: I0930 15:01:58.034909 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6d689559c5-h7fmf_e047fa81-9835-4bca-95c4-c8d24f97857b/nmstate-webhook/0.log" Sep 30 15:02:10 crc kubenswrapper[4840]: I0930 15:02:10.117209 4840 scope.go:117] "RemoveContainer" containerID="409430684b16cc7e5d86eb8dbc5c06065665c2c71a8402382eb805da2eca622c" Sep 30 15:02:10 crc kubenswrapper[4840]: E0930 15:02:10.118037 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 15:02:12 crc kubenswrapper[4840]: I0930 15:02:12.347500 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5d688f5ffc-nv2dz_e2682a09-9f63-42a1-882a-368be12b8be0/kube-rbac-proxy/0.log" Sep 30 15:02:12 crc kubenswrapper[4840]: I0930 15:02:12.507867 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5d688f5ffc-nv2dz_e2682a09-9f63-42a1-882a-368be12b8be0/controller/0.log" Sep 30 15:02:12 crc kubenswrapper[4840]: I0930 15:02:12.572518 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sqx2j_624a00a9-8fda-4e40-9df0-e8452ed89d28/cp-frr-files/0.log" Sep 30 15:02:12 crc kubenswrapper[4840]: I0930 15:02:12.685672 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sqx2j_624a00a9-8fda-4e40-9df0-e8452ed89d28/cp-frr-files/0.log" Sep 30 15:02:12 crc kubenswrapper[4840]: I0930 15:02:12.709225 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sqx2j_624a00a9-8fda-4e40-9df0-e8452ed89d28/cp-reloader/0.log" Sep 30 15:02:12 crc kubenswrapper[4840]: I0930 15:02:12.720185 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sqx2j_624a00a9-8fda-4e40-9df0-e8452ed89d28/cp-metrics/0.log" Sep 30 15:02:12 crc kubenswrapper[4840]: I0930 15:02:12.752143 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sqx2j_624a00a9-8fda-4e40-9df0-e8452ed89d28/cp-reloader/0.log" Sep 30 15:02:12 crc kubenswrapper[4840]: I0930 15:02:12.913304 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sqx2j_624a00a9-8fda-4e40-9df0-e8452ed89d28/cp-frr-files/0.log" Sep 30 15:02:12 crc kubenswrapper[4840]: I0930 15:02:12.928218 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sqx2j_624a00a9-8fda-4e40-9df0-e8452ed89d28/cp-reloader/0.log" Sep 30 15:02:12 crc kubenswrapper[4840]: I0930 15:02:12.940760 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sqx2j_624a00a9-8fda-4e40-9df0-e8452ed89d28/cp-metrics/0.log" Sep 30 15:02:12 crc kubenswrapper[4840]: I0930 15:02:12.957452 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sqx2j_624a00a9-8fda-4e40-9df0-e8452ed89d28/cp-metrics/0.log" Sep 30 15:02:13 crc kubenswrapper[4840]: I0930 15:02:13.106374 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sqx2j_624a00a9-8fda-4e40-9df0-e8452ed89d28/cp-reloader/0.log" Sep 30 15:02:13 crc kubenswrapper[4840]: I0930 15:02:13.106397 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sqx2j_624a00a9-8fda-4e40-9df0-e8452ed89d28/cp-frr-files/0.log" Sep 30 15:02:13 crc kubenswrapper[4840]: I0930 15:02:13.114705 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sqx2j_624a00a9-8fda-4e40-9df0-e8452ed89d28/cp-metrics/0.log" Sep 30 15:02:13 crc kubenswrapper[4840]: I0930 15:02:13.120765 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sqx2j_624a00a9-8fda-4e40-9df0-e8452ed89d28/controller/0.log" Sep 30 15:02:13 crc kubenswrapper[4840]: I0930 15:02:13.281672 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sqx2j_624a00a9-8fda-4e40-9df0-e8452ed89d28/kube-rbac-proxy/0.log" Sep 30 15:02:13 crc kubenswrapper[4840]: I0930 15:02:13.289776 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sqx2j_624a00a9-8fda-4e40-9df0-e8452ed89d28/frr-metrics/0.log" Sep 30 15:02:13 crc kubenswrapper[4840]: I0930 15:02:13.306785 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sqx2j_624a00a9-8fda-4e40-9df0-e8452ed89d28/kube-rbac-proxy-frr/0.log" Sep 30 15:02:13 crc kubenswrapper[4840]: I0930 15:02:13.472149 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sqx2j_624a00a9-8fda-4e40-9df0-e8452ed89d28/reloader/0.log" Sep 30 15:02:13 crc kubenswrapper[4840]: I0930 15:02:13.556978 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-5478bdb765-8qmmm_7348a607-eb78-4fbf-a620-9be8180b8503/frr-k8s-webhook-server/0.log" Sep 30 15:02:13 crc kubenswrapper[4840]: I0930 15:02:13.680241 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6c8fd557c5-29r6x_ec2b2b7b-3b95-42c0-bfde-d8206541fe62/manager/0.log" Sep 30 15:02:13 crc kubenswrapper[4840]: I0930 15:02:13.906923 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-79c987c9f-v9dzf_1e99fa8d-4919-483b-a71e-ee137198e55c/webhook-server/0.log" Sep 30 15:02:14 crc kubenswrapper[4840]: I0930 15:02:14.027510 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-hnv4n_a15e1237-7226-4180-8c27-fb06dffbd2d7/kube-rbac-proxy/0.log" Sep 30 15:02:14 crc kubenswrapper[4840]: I0930 15:02:14.564747 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-hnv4n_a15e1237-7226-4180-8c27-fb06dffbd2d7/speaker/0.log" Sep 30 15:02:14 crc kubenswrapper[4840]: I0930 15:02:14.716456 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sqx2j_624a00a9-8fda-4e40-9df0-e8452ed89d28/frr/0.log" Sep 30 15:02:21 crc kubenswrapper[4840]: I0930 15:02:21.117875 4840 scope.go:117] "RemoveContainer" containerID="409430684b16cc7e5d86eb8dbc5c06065665c2c71a8402382eb805da2eca622c" Sep 30 15:02:21 crc kubenswrapper[4840]: E0930 15:02:21.118663 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 15:02:26 crc kubenswrapper[4840]: I0930 15:02:26.518178 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6_0baa1415-f354-45dd-a352-140902beba82/util/0.log" Sep 30 15:02:26 crc kubenswrapper[4840]: I0930 15:02:26.720621 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6_0baa1415-f354-45dd-a352-140902beba82/util/0.log" Sep 30 15:02:26 crc kubenswrapper[4840]: I0930 15:02:26.729876 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6_0baa1415-f354-45dd-a352-140902beba82/pull/0.log" Sep 30 15:02:26 crc kubenswrapper[4840]: I0930 15:02:26.764736 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6_0baa1415-f354-45dd-a352-140902beba82/pull/0.log" Sep 30 15:02:26 crc kubenswrapper[4840]: I0930 15:02:26.904115 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6_0baa1415-f354-45dd-a352-140902beba82/util/0.log" Sep 30 15:02:26 crc kubenswrapper[4840]: I0930 15:02:26.940507 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6_0baa1415-f354-45dd-a352-140902beba82/extract/0.log" Sep 30 15:02:26 crc kubenswrapper[4840]: I0930 15:02:26.951279 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcls7f6_0baa1415-f354-45dd-a352-140902beba82/pull/0.log" Sep 30 15:02:27 crc kubenswrapper[4840]: I0930 15:02:27.104613 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zbb8b_233f29d4-5cbe-4cd1-b8ac-e40cf332d381/extract-utilities/0.log" Sep 30 15:02:27 crc kubenswrapper[4840]: I0930 15:02:27.264816 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zbb8b_233f29d4-5cbe-4cd1-b8ac-e40cf332d381/extract-utilities/0.log" Sep 30 15:02:27 crc kubenswrapper[4840]: I0930 15:02:27.284124 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zbb8b_233f29d4-5cbe-4cd1-b8ac-e40cf332d381/extract-content/0.log" Sep 30 15:02:27 crc kubenswrapper[4840]: I0930 15:02:27.310160 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zbb8b_233f29d4-5cbe-4cd1-b8ac-e40cf332d381/extract-content/0.log" Sep 30 15:02:27 crc kubenswrapper[4840]: I0930 15:02:27.426514 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zbb8b_233f29d4-5cbe-4cd1-b8ac-e40cf332d381/extract-utilities/0.log" Sep 30 15:02:27 crc kubenswrapper[4840]: I0930 15:02:27.438584 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zbb8b_233f29d4-5cbe-4cd1-b8ac-e40cf332d381/extract-content/0.log" Sep 30 15:02:27 crc kubenswrapper[4840]: I0930 15:02:27.664738 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-t2tp4_e5a51075-4798-48bf-ac82-67a642d1269a/extract-utilities/0.log" Sep 30 15:02:27 crc kubenswrapper[4840]: I0930 15:02:27.918967 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-t2tp4_e5a51075-4798-48bf-ac82-67a642d1269a/extract-content/0.log" Sep 30 15:02:27 crc kubenswrapper[4840]: I0930 15:02:27.919537 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-t2tp4_e5a51075-4798-48bf-ac82-67a642d1269a/extract-content/0.log" Sep 30 15:02:27 crc kubenswrapper[4840]: I0930 15:02:27.948988 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-t2tp4_e5a51075-4798-48bf-ac82-67a642d1269a/extract-utilities/0.log" Sep 30 15:02:28 crc kubenswrapper[4840]: I0930 15:02:28.113944 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zbb8b_233f29d4-5cbe-4cd1-b8ac-e40cf332d381/registry-server/0.log" Sep 30 15:02:28 crc kubenswrapper[4840]: I0930 15:02:28.163881 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-t2tp4_e5a51075-4798-48bf-ac82-67a642d1269a/extract-utilities/0.log" Sep 30 15:02:28 crc kubenswrapper[4840]: I0930 15:02:28.175924 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-t2tp4_e5a51075-4798-48bf-ac82-67a642d1269a/extract-content/0.log" Sep 30 15:02:28 crc kubenswrapper[4840]: I0930 15:02:28.361697 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j_e8ac07f8-b0b5-4ff0-bae8-4d6594567c32/util/0.log" Sep 30 15:02:28 crc kubenswrapper[4840]: I0930 15:02:28.597955 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j_e8ac07f8-b0b5-4ff0-bae8-4d6594567c32/util/0.log" Sep 30 15:02:28 crc kubenswrapper[4840]: I0930 15:02:28.663794 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j_e8ac07f8-b0b5-4ff0-bae8-4d6594567c32/pull/0.log" Sep 30 15:02:28 crc kubenswrapper[4840]: I0930 15:02:28.703945 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j_e8ac07f8-b0b5-4ff0-bae8-4d6594567c32/pull/0.log" Sep 30 15:02:28 crc kubenswrapper[4840]: I0930 15:02:28.733747 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-t2tp4_e5a51075-4798-48bf-ac82-67a642d1269a/registry-server/0.log" Sep 30 15:02:28 crc kubenswrapper[4840]: I0930 15:02:28.814140 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j_e8ac07f8-b0b5-4ff0-bae8-4d6594567c32/util/0.log" Sep 30 15:02:28 crc kubenswrapper[4840]: I0930 15:02:28.880153 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j_e8ac07f8-b0b5-4ff0-bae8-4d6594567c32/pull/0.log" Sep 30 15:02:28 crc kubenswrapper[4840]: I0930 15:02:28.880301 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96skc6j_e8ac07f8-b0b5-4ff0-bae8-4d6594567c32/extract/0.log" Sep 30 15:02:29 crc kubenswrapper[4840]: I0930 15:02:29.021748 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-hfdnb_b0c00d91-ae86-48df-84b1-1e1c29b7ff38/marketplace-operator/0.log" Sep 30 15:02:29 crc kubenswrapper[4840]: I0930 15:02:29.075622 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f85rt_0207a121-1290-431e-85a5-e5bfc5355d50/extract-utilities/0.log" Sep 30 15:02:29 crc kubenswrapper[4840]: I0930 15:02:29.622903 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f85rt_0207a121-1290-431e-85a5-e5bfc5355d50/extract-utilities/0.log" Sep 30 15:02:29 crc kubenswrapper[4840]: I0930 15:02:29.625129 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f85rt_0207a121-1290-431e-85a5-e5bfc5355d50/extract-content/0.log" Sep 30 15:02:29 crc kubenswrapper[4840]: I0930 15:02:29.651293 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f85rt_0207a121-1290-431e-85a5-e5bfc5355d50/extract-content/0.log" Sep 30 15:02:29 crc kubenswrapper[4840]: I0930 15:02:29.826801 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f85rt_0207a121-1290-431e-85a5-e5bfc5355d50/extract-utilities/0.log" Sep 30 15:02:29 crc kubenswrapper[4840]: I0930 15:02:29.832355 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f85rt_0207a121-1290-431e-85a5-e5bfc5355d50/extract-content/0.log" Sep 30 15:02:30 crc kubenswrapper[4840]: I0930 15:02:30.028437 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f85rt_0207a121-1290-431e-85a5-e5bfc5355d50/registry-server/0.log" Sep 30 15:02:30 crc kubenswrapper[4840]: I0930 15:02:30.043609 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rzkrh_0ae5abcf-aeb1-4607-8285-cd6ea1d1f912/extract-utilities/0.log" Sep 30 15:02:30 crc kubenswrapper[4840]: I0930 15:02:30.215714 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rzkrh_0ae5abcf-aeb1-4607-8285-cd6ea1d1f912/extract-utilities/0.log" Sep 30 15:02:30 crc kubenswrapper[4840]: I0930 15:02:30.247228 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rzkrh_0ae5abcf-aeb1-4607-8285-cd6ea1d1f912/extract-content/0.log" Sep 30 15:02:30 crc kubenswrapper[4840]: I0930 15:02:30.254141 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rzkrh_0ae5abcf-aeb1-4607-8285-cd6ea1d1f912/extract-content/0.log" Sep 30 15:02:30 crc kubenswrapper[4840]: I0930 15:02:30.397029 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rzkrh_0ae5abcf-aeb1-4607-8285-cd6ea1d1f912/extract-utilities/0.log" Sep 30 15:02:30 crc kubenswrapper[4840]: I0930 15:02:30.397328 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rzkrh_0ae5abcf-aeb1-4607-8285-cd6ea1d1f912/extract-content/0.log" Sep 30 15:02:30 crc kubenswrapper[4840]: I0930 15:02:30.807175 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rzkrh_0ae5abcf-aeb1-4607-8285-cd6ea1d1f912/registry-server/0.log" Sep 30 15:02:32 crc kubenswrapper[4840]: I0930 15:02:32.125116 4840 scope.go:117] "RemoveContainer" containerID="409430684b16cc7e5d86eb8dbc5c06065665c2c71a8402382eb805da2eca622c" Sep 30 15:02:32 crc kubenswrapper[4840]: E0930 15:02:32.125710 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 15:02:47 crc kubenswrapper[4840]: I0930 15:02:47.116279 4840 scope.go:117] "RemoveContainer" containerID="409430684b16cc7e5d86eb8dbc5c06065665c2c71a8402382eb805da2eca622c" Sep 30 15:02:47 crc kubenswrapper[4840]: E0930 15:02:47.116888 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 15:03:02 crc kubenswrapper[4840]: I0930 15:03:02.129383 4840 scope.go:117] "RemoveContainer" containerID="409430684b16cc7e5d86eb8dbc5c06065665c2c71a8402382eb805da2eca622c" Sep 30 15:03:02 crc kubenswrapper[4840]: E0930 15:03:02.130216 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 15:03:15 crc kubenswrapper[4840]: I0930 15:03:15.116784 4840 scope.go:117] "RemoveContainer" containerID="409430684b16cc7e5d86eb8dbc5c06065665c2c71a8402382eb805da2eca622c" Sep 30 15:03:15 crc kubenswrapper[4840]: E0930 15:03:15.117670 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 15:03:29 crc kubenswrapper[4840]: I0930 15:03:29.116148 4840 scope.go:117] "RemoveContainer" containerID="409430684b16cc7e5d86eb8dbc5c06065665c2c71a8402382eb805da2eca622c" Sep 30 15:03:29 crc kubenswrapper[4840]: E0930 15:03:29.117203 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 15:03:43 crc kubenswrapper[4840]: I0930 15:03:43.116812 4840 scope.go:117] "RemoveContainer" containerID="409430684b16cc7e5d86eb8dbc5c06065665c2c71a8402382eb805da2eca622c" Sep 30 15:03:43 crc kubenswrapper[4840]: E0930 15:03:43.117729 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 15:03:54 crc kubenswrapper[4840]: I0930 15:03:54.116787 4840 scope.go:117] "RemoveContainer" containerID="409430684b16cc7e5d86eb8dbc5c06065665c2c71a8402382eb805da2eca622c" Sep 30 15:03:54 crc kubenswrapper[4840]: E0930 15:03:54.117605 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 15:04:08 crc kubenswrapper[4840]: I0930 15:04:08.116189 4840 scope.go:117] "RemoveContainer" containerID="409430684b16cc7e5d86eb8dbc5c06065665c2c71a8402382eb805da2eca622c" Sep 30 15:04:08 crc kubenswrapper[4840]: E0930 15:04:08.116758 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 15:04:23 crc kubenswrapper[4840]: I0930 15:04:23.116219 4840 scope.go:117] "RemoveContainer" containerID="409430684b16cc7e5d86eb8dbc5c06065665c2c71a8402382eb805da2eca622c" Sep 30 15:04:23 crc kubenswrapper[4840]: E0930 15:04:23.118259 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 15:04:28 crc kubenswrapper[4840]: I0930 15:04:28.394623 4840 generic.go:334] "Generic (PLEG): container finished" podID="16639026-05f0-4186-9f32-2b9c7e52bff1" containerID="a46d157e054e258dc85184bcd87f7f9cd75929c160435b559b14641d1c2986c1" exitCode=0 Sep 30 15:04:28 crc kubenswrapper[4840]: I0930 15:04:28.394781 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gf4gj/must-gather-2hpk7" event={"ID":"16639026-05f0-4186-9f32-2b9c7e52bff1","Type":"ContainerDied","Data":"a46d157e054e258dc85184bcd87f7f9cd75929c160435b559b14641d1c2986c1"} Sep 30 15:04:28 crc kubenswrapper[4840]: I0930 15:04:28.395972 4840 scope.go:117] "RemoveContainer" containerID="a46d157e054e258dc85184bcd87f7f9cd75929c160435b559b14641d1c2986c1" Sep 30 15:04:28 crc kubenswrapper[4840]: I0930 15:04:28.969048 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gf4gj_must-gather-2hpk7_16639026-05f0-4186-9f32-2b9c7e52bff1/gather/0.log" Sep 30 15:04:37 crc kubenswrapper[4840]: I0930 15:04:37.119330 4840 scope.go:117] "RemoveContainer" containerID="409430684b16cc7e5d86eb8dbc5c06065665c2c71a8402382eb805da2eca622c" Sep 30 15:04:37 crc kubenswrapper[4840]: E0930 15:04:37.120960 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 15:04:40 crc kubenswrapper[4840]: I0930 15:04:40.787456 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-gf4gj/must-gather-2hpk7"] Sep 30 15:04:40 crc kubenswrapper[4840]: I0930 15:04:40.789277 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-gf4gj/must-gather-2hpk7" podUID="16639026-05f0-4186-9f32-2b9c7e52bff1" containerName="copy" containerID="cri-o://157e02fdc66df15d3fc935e6af798167c678a29e75c7302f22f037cb92927f7c" gracePeriod=2 Sep 30 15:04:40 crc kubenswrapper[4840]: I0930 15:04:40.796456 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-gf4gj/must-gather-2hpk7"] Sep 30 15:04:41 crc kubenswrapper[4840]: I0930 15:04:41.248799 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gf4gj_must-gather-2hpk7_16639026-05f0-4186-9f32-2b9c7e52bff1/copy/0.log" Sep 30 15:04:41 crc kubenswrapper[4840]: I0930 15:04:41.249386 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gf4gj/must-gather-2hpk7" Sep 30 15:04:41 crc kubenswrapper[4840]: I0930 15:04:41.333340 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fhxbw\" (UniqueName: \"kubernetes.io/projected/16639026-05f0-4186-9f32-2b9c7e52bff1-kube-api-access-fhxbw\") pod \"16639026-05f0-4186-9f32-2b9c7e52bff1\" (UID: \"16639026-05f0-4186-9f32-2b9c7e52bff1\") " Sep 30 15:04:41 crc kubenswrapper[4840]: I0930 15:04:41.333747 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/16639026-05f0-4186-9f32-2b9c7e52bff1-must-gather-output\") pod \"16639026-05f0-4186-9f32-2b9c7e52bff1\" (UID: \"16639026-05f0-4186-9f32-2b9c7e52bff1\") " Sep 30 15:04:41 crc kubenswrapper[4840]: I0930 15:04:41.339477 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16639026-05f0-4186-9f32-2b9c7e52bff1-kube-api-access-fhxbw" (OuterVolumeSpecName: "kube-api-access-fhxbw") pod "16639026-05f0-4186-9f32-2b9c7e52bff1" (UID: "16639026-05f0-4186-9f32-2b9c7e52bff1"). InnerVolumeSpecName "kube-api-access-fhxbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:04:41 crc kubenswrapper[4840]: I0930 15:04:41.436268 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fhxbw\" (UniqueName: \"kubernetes.io/projected/16639026-05f0-4186-9f32-2b9c7e52bff1-kube-api-access-fhxbw\") on node \"crc\" DevicePath \"\"" Sep 30 15:04:41 crc kubenswrapper[4840]: I0930 15:04:41.486493 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16639026-05f0-4186-9f32-2b9c7e52bff1-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "16639026-05f0-4186-9f32-2b9c7e52bff1" (UID: "16639026-05f0-4186-9f32-2b9c7e52bff1"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:04:41 crc kubenswrapper[4840]: I0930 15:04:41.528691 4840 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gf4gj_must-gather-2hpk7_16639026-05f0-4186-9f32-2b9c7e52bff1/copy/0.log" Sep 30 15:04:41 crc kubenswrapper[4840]: I0930 15:04:41.529242 4840 generic.go:334] "Generic (PLEG): container finished" podID="16639026-05f0-4186-9f32-2b9c7e52bff1" containerID="157e02fdc66df15d3fc935e6af798167c678a29e75c7302f22f037cb92927f7c" exitCode=143 Sep 30 15:04:41 crc kubenswrapper[4840]: I0930 15:04:41.529297 4840 scope.go:117] "RemoveContainer" containerID="157e02fdc66df15d3fc935e6af798167c678a29e75c7302f22f037cb92927f7c" Sep 30 15:04:41 crc kubenswrapper[4840]: I0930 15:04:41.529300 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gf4gj/must-gather-2hpk7" Sep 30 15:04:41 crc kubenswrapper[4840]: I0930 15:04:41.537632 4840 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/16639026-05f0-4186-9f32-2b9c7e52bff1-must-gather-output\") on node \"crc\" DevicePath \"\"" Sep 30 15:04:41 crc kubenswrapper[4840]: I0930 15:04:41.547831 4840 scope.go:117] "RemoveContainer" containerID="a46d157e054e258dc85184bcd87f7f9cd75929c160435b559b14641d1c2986c1" Sep 30 15:04:41 crc kubenswrapper[4840]: I0930 15:04:41.612902 4840 scope.go:117] "RemoveContainer" containerID="157e02fdc66df15d3fc935e6af798167c678a29e75c7302f22f037cb92927f7c" Sep 30 15:04:41 crc kubenswrapper[4840]: E0930 15:04:41.614109 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"157e02fdc66df15d3fc935e6af798167c678a29e75c7302f22f037cb92927f7c\": container with ID starting with 157e02fdc66df15d3fc935e6af798167c678a29e75c7302f22f037cb92927f7c not found: ID does not exist" containerID="157e02fdc66df15d3fc935e6af798167c678a29e75c7302f22f037cb92927f7c" Sep 30 15:04:41 crc kubenswrapper[4840]: I0930 15:04:41.614145 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"157e02fdc66df15d3fc935e6af798167c678a29e75c7302f22f037cb92927f7c"} err="failed to get container status \"157e02fdc66df15d3fc935e6af798167c678a29e75c7302f22f037cb92927f7c\": rpc error: code = NotFound desc = could not find container \"157e02fdc66df15d3fc935e6af798167c678a29e75c7302f22f037cb92927f7c\": container with ID starting with 157e02fdc66df15d3fc935e6af798167c678a29e75c7302f22f037cb92927f7c not found: ID does not exist" Sep 30 15:04:41 crc kubenswrapper[4840]: I0930 15:04:41.614165 4840 scope.go:117] "RemoveContainer" containerID="a46d157e054e258dc85184bcd87f7f9cd75929c160435b559b14641d1c2986c1" Sep 30 15:04:41 crc kubenswrapper[4840]: E0930 15:04:41.614493 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a46d157e054e258dc85184bcd87f7f9cd75929c160435b559b14641d1c2986c1\": container with ID starting with a46d157e054e258dc85184bcd87f7f9cd75929c160435b559b14641d1c2986c1 not found: ID does not exist" containerID="a46d157e054e258dc85184bcd87f7f9cd75929c160435b559b14641d1c2986c1" Sep 30 15:04:41 crc kubenswrapper[4840]: I0930 15:04:41.614543 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a46d157e054e258dc85184bcd87f7f9cd75929c160435b559b14641d1c2986c1"} err="failed to get container status \"a46d157e054e258dc85184bcd87f7f9cd75929c160435b559b14641d1c2986c1\": rpc error: code = NotFound desc = could not find container \"a46d157e054e258dc85184bcd87f7f9cd75929c160435b559b14641d1c2986c1\": container with ID starting with a46d157e054e258dc85184bcd87f7f9cd75929c160435b559b14641d1c2986c1 not found: ID does not exist" Sep 30 15:04:42 crc kubenswrapper[4840]: I0930 15:04:42.126168 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16639026-05f0-4186-9f32-2b9c7e52bff1" path="/var/lib/kubelet/pods/16639026-05f0-4186-9f32-2b9c7e52bff1/volumes" Sep 30 15:04:50 crc kubenswrapper[4840]: I0930 15:04:50.116766 4840 scope.go:117] "RemoveContainer" containerID="409430684b16cc7e5d86eb8dbc5c06065665c2c71a8402382eb805da2eca622c" Sep 30 15:04:50 crc kubenswrapper[4840]: E0930 15:04:50.117545 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 15:05:02 crc kubenswrapper[4840]: I0930 15:05:02.858750 4840 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-l85xl"] Sep 30 15:05:02 crc kubenswrapper[4840]: E0930 15:05:02.859778 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16639026-05f0-4186-9f32-2b9c7e52bff1" containerName="gather" Sep 30 15:05:02 crc kubenswrapper[4840]: I0930 15:05:02.859795 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="16639026-05f0-4186-9f32-2b9c7e52bff1" containerName="gather" Sep 30 15:05:02 crc kubenswrapper[4840]: E0930 15:05:02.859845 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a7d4b45-82ea-4041-8391-f96864afea7a" containerName="extract-content" Sep 30 15:05:02 crc kubenswrapper[4840]: I0930 15:05:02.859853 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a7d4b45-82ea-4041-8391-f96864afea7a" containerName="extract-content" Sep 30 15:05:02 crc kubenswrapper[4840]: E0930 15:05:02.859872 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16639026-05f0-4186-9f32-2b9c7e52bff1" containerName="copy" Sep 30 15:05:02 crc kubenswrapper[4840]: I0930 15:05:02.859879 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="16639026-05f0-4186-9f32-2b9c7e52bff1" containerName="copy" Sep 30 15:05:02 crc kubenswrapper[4840]: E0930 15:05:02.859892 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a7d4b45-82ea-4041-8391-f96864afea7a" containerName="extract-utilities" Sep 30 15:05:02 crc kubenswrapper[4840]: I0930 15:05:02.859903 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a7d4b45-82ea-4041-8391-f96864afea7a" containerName="extract-utilities" Sep 30 15:05:02 crc kubenswrapper[4840]: E0930 15:05:02.859914 4840 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a7d4b45-82ea-4041-8391-f96864afea7a" containerName="registry-server" Sep 30 15:05:02 crc kubenswrapper[4840]: I0930 15:05:02.859922 4840 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a7d4b45-82ea-4041-8391-f96864afea7a" containerName="registry-server" Sep 30 15:05:02 crc kubenswrapper[4840]: I0930 15:05:02.860124 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a7d4b45-82ea-4041-8391-f96864afea7a" containerName="registry-server" Sep 30 15:05:02 crc kubenswrapper[4840]: I0930 15:05:02.860145 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="16639026-05f0-4186-9f32-2b9c7e52bff1" containerName="copy" Sep 30 15:05:02 crc kubenswrapper[4840]: I0930 15:05:02.860163 4840 memory_manager.go:354] "RemoveStaleState removing state" podUID="16639026-05f0-4186-9f32-2b9c7e52bff1" containerName="gather" Sep 30 15:05:02 crc kubenswrapper[4840]: I0930 15:05:02.861981 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l85xl" Sep 30 15:05:02 crc kubenswrapper[4840]: I0930 15:05:02.882413 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l85xl"] Sep 30 15:05:02 crc kubenswrapper[4840]: I0930 15:05:02.933107 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/435addf3-8422-4700-baf4-0ac5f1ccd41e-catalog-content\") pod \"community-operators-l85xl\" (UID: \"435addf3-8422-4700-baf4-0ac5f1ccd41e\") " pod="openshift-marketplace/community-operators-l85xl" Sep 30 15:05:02 crc kubenswrapper[4840]: I0930 15:05:02.933459 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5h8m\" (UniqueName: \"kubernetes.io/projected/435addf3-8422-4700-baf4-0ac5f1ccd41e-kube-api-access-p5h8m\") pod \"community-operators-l85xl\" (UID: \"435addf3-8422-4700-baf4-0ac5f1ccd41e\") " pod="openshift-marketplace/community-operators-l85xl" Sep 30 15:05:02 crc kubenswrapper[4840]: I0930 15:05:02.933712 4840 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/435addf3-8422-4700-baf4-0ac5f1ccd41e-utilities\") pod \"community-operators-l85xl\" (UID: \"435addf3-8422-4700-baf4-0ac5f1ccd41e\") " pod="openshift-marketplace/community-operators-l85xl" Sep 30 15:05:03 crc kubenswrapper[4840]: I0930 15:05:03.035909 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/435addf3-8422-4700-baf4-0ac5f1ccd41e-utilities\") pod \"community-operators-l85xl\" (UID: \"435addf3-8422-4700-baf4-0ac5f1ccd41e\") " pod="openshift-marketplace/community-operators-l85xl" Sep 30 15:05:03 crc kubenswrapper[4840]: I0930 15:05:03.036164 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/435addf3-8422-4700-baf4-0ac5f1ccd41e-catalog-content\") pod \"community-operators-l85xl\" (UID: \"435addf3-8422-4700-baf4-0ac5f1ccd41e\") " pod="openshift-marketplace/community-operators-l85xl" Sep 30 15:05:03 crc kubenswrapper[4840]: I0930 15:05:03.036302 4840 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5h8m\" (UniqueName: \"kubernetes.io/projected/435addf3-8422-4700-baf4-0ac5f1ccd41e-kube-api-access-p5h8m\") pod \"community-operators-l85xl\" (UID: \"435addf3-8422-4700-baf4-0ac5f1ccd41e\") " pod="openshift-marketplace/community-operators-l85xl" Sep 30 15:05:03 crc kubenswrapper[4840]: I0930 15:05:03.037346 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/435addf3-8422-4700-baf4-0ac5f1ccd41e-utilities\") pod \"community-operators-l85xl\" (UID: \"435addf3-8422-4700-baf4-0ac5f1ccd41e\") " pod="openshift-marketplace/community-operators-l85xl" Sep 30 15:05:03 crc kubenswrapper[4840]: I0930 15:05:03.037466 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/435addf3-8422-4700-baf4-0ac5f1ccd41e-catalog-content\") pod \"community-operators-l85xl\" (UID: \"435addf3-8422-4700-baf4-0ac5f1ccd41e\") " pod="openshift-marketplace/community-operators-l85xl" Sep 30 15:05:03 crc kubenswrapper[4840]: I0930 15:05:03.056619 4840 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5h8m\" (UniqueName: \"kubernetes.io/projected/435addf3-8422-4700-baf4-0ac5f1ccd41e-kube-api-access-p5h8m\") pod \"community-operators-l85xl\" (UID: \"435addf3-8422-4700-baf4-0ac5f1ccd41e\") " pod="openshift-marketplace/community-operators-l85xl" Sep 30 15:05:03 crc kubenswrapper[4840]: I0930 15:05:03.116794 4840 scope.go:117] "RemoveContainer" containerID="409430684b16cc7e5d86eb8dbc5c06065665c2c71a8402382eb805da2eca622c" Sep 30 15:05:03 crc kubenswrapper[4840]: E0930 15:05:03.117112 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 15:05:03 crc kubenswrapper[4840]: I0930 15:05:03.186600 4840 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l85xl" Sep 30 15:05:03 crc kubenswrapper[4840]: I0930 15:05:03.761149 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l85xl"] Sep 30 15:05:04 crc kubenswrapper[4840]: I0930 15:05:04.722842 4840 generic.go:334] "Generic (PLEG): container finished" podID="435addf3-8422-4700-baf4-0ac5f1ccd41e" containerID="272e7ca52ee67c5aa3e04a63c74ce7c366957b02ef45882b7ee5f57a0c405fae" exitCode=0 Sep 30 15:05:04 crc kubenswrapper[4840]: I0930 15:05:04.722891 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l85xl" event={"ID":"435addf3-8422-4700-baf4-0ac5f1ccd41e","Type":"ContainerDied","Data":"272e7ca52ee67c5aa3e04a63c74ce7c366957b02ef45882b7ee5f57a0c405fae"} Sep 30 15:05:04 crc kubenswrapper[4840]: I0930 15:05:04.722916 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l85xl" event={"ID":"435addf3-8422-4700-baf4-0ac5f1ccd41e","Type":"ContainerStarted","Data":"dae25e84f72af99dde8d5cbca9381545dc984b53c7fe7f5eef0004b23e6d28eb"} Sep 30 15:05:04 crc kubenswrapper[4840]: I0930 15:05:04.726426 4840 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 15:05:07 crc kubenswrapper[4840]: I0930 15:05:07.752779 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l85xl" event={"ID":"435addf3-8422-4700-baf4-0ac5f1ccd41e","Type":"ContainerStarted","Data":"c58fabe653e8ece886c222ae330d1f924478e09db6a80aa3c812482ac055f508"} Sep 30 15:05:08 crc kubenswrapper[4840]: I0930 15:05:08.764054 4840 generic.go:334] "Generic (PLEG): container finished" podID="435addf3-8422-4700-baf4-0ac5f1ccd41e" containerID="c58fabe653e8ece886c222ae330d1f924478e09db6a80aa3c812482ac055f508" exitCode=0 Sep 30 15:05:08 crc kubenswrapper[4840]: I0930 15:05:08.764104 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l85xl" event={"ID":"435addf3-8422-4700-baf4-0ac5f1ccd41e","Type":"ContainerDied","Data":"c58fabe653e8ece886c222ae330d1f924478e09db6a80aa3c812482ac055f508"} Sep 30 15:05:09 crc kubenswrapper[4840]: I0930 15:05:09.775979 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l85xl" event={"ID":"435addf3-8422-4700-baf4-0ac5f1ccd41e","Type":"ContainerStarted","Data":"6edd7f5e815f23e8ee4090b3eefc8d9e8e41cfb9190dfc85ea5dae7262d2784a"} Sep 30 15:05:09 crc kubenswrapper[4840]: I0930 15:05:09.798696 4840 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-l85xl" podStartSLOduration=3.282673808 podStartE2EDuration="7.798680418s" podCreationTimestamp="2025-09-30 15:05:02 +0000 UTC" firstStartedPulling="2025-09-30 15:05:04.726215167 +0000 UTC m=+4133.355301590" lastFinishedPulling="2025-09-30 15:05:09.242221757 +0000 UTC m=+4137.871308200" observedRunningTime="2025-09-30 15:05:09.79295389 +0000 UTC m=+4138.422040323" watchObservedRunningTime="2025-09-30 15:05:09.798680418 +0000 UTC m=+4138.427766841" Sep 30 15:05:13 crc kubenswrapper[4840]: I0930 15:05:13.187401 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-l85xl" Sep 30 15:05:13 crc kubenswrapper[4840]: I0930 15:05:13.188076 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-l85xl" Sep 30 15:05:13 crc kubenswrapper[4840]: I0930 15:05:13.242782 4840 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-l85xl" Sep 30 15:05:14 crc kubenswrapper[4840]: I0930 15:05:14.116280 4840 scope.go:117] "RemoveContainer" containerID="409430684b16cc7e5d86eb8dbc5c06065665c2c71a8402382eb805da2eca622c" Sep 30 15:05:14 crc kubenswrapper[4840]: E0930 15:05:14.116790 4840 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-747gk_openshift-machine-config-operator(10e8b890-7f20-4a36-8e03-898620cf599a)\"" pod="openshift-machine-config-operator/machine-config-daemon-747gk" podUID="10e8b890-7f20-4a36-8e03-898620cf599a" Sep 30 15:05:23 crc kubenswrapper[4840]: I0930 15:05:23.241625 4840 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-l85xl" Sep 30 15:05:23 crc kubenswrapper[4840]: I0930 15:05:23.312245 4840 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l85xl"] Sep 30 15:05:23 crc kubenswrapper[4840]: I0930 15:05:23.340075 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t2tp4"] Sep 30 15:05:23 crc kubenswrapper[4840]: I0930 15:05:23.340303 4840 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-t2tp4" podUID="e5a51075-4798-48bf-ac82-67a642d1269a" containerName="registry-server" containerID="cri-o://3210f4d8b25c37367beb5eb6807e2586568932e60d098eaa2175ae7fb13b0a98" gracePeriod=2 Sep 30 15:05:23 crc kubenswrapper[4840]: I0930 15:05:23.793365 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t2tp4" Sep 30 15:05:23 crc kubenswrapper[4840]: I0930 15:05:23.886867 4840 generic.go:334] "Generic (PLEG): container finished" podID="e5a51075-4798-48bf-ac82-67a642d1269a" containerID="3210f4d8b25c37367beb5eb6807e2586568932e60d098eaa2175ae7fb13b0a98" exitCode=0 Sep 30 15:05:23 crc kubenswrapper[4840]: I0930 15:05:23.887178 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t2tp4" event={"ID":"e5a51075-4798-48bf-ac82-67a642d1269a","Type":"ContainerDied","Data":"3210f4d8b25c37367beb5eb6807e2586568932e60d098eaa2175ae7fb13b0a98"} Sep 30 15:05:23 crc kubenswrapper[4840]: I0930 15:05:23.887216 4840 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t2tp4" Sep 30 15:05:23 crc kubenswrapper[4840]: I0930 15:05:23.887224 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t2tp4" event={"ID":"e5a51075-4798-48bf-ac82-67a642d1269a","Type":"ContainerDied","Data":"9759a8a21a3167dffe576b212f26fb1883691c78b2390d18f06d5dec6ea43cd1"} Sep 30 15:05:23 crc kubenswrapper[4840]: I0930 15:05:23.887264 4840 scope.go:117] "RemoveContainer" containerID="3210f4d8b25c37367beb5eb6807e2586568932e60d098eaa2175ae7fb13b0a98" Sep 30 15:05:23 crc kubenswrapper[4840]: I0930 15:05:23.923894 4840 scope.go:117] "RemoveContainer" containerID="8f002688f13a0db9631486a4b1602411ec8f5de05c142f1a40741b881204a71c" Sep 30 15:05:23 crc kubenswrapper[4840]: I0930 15:05:23.947437 4840 scope.go:117] "RemoveContainer" containerID="c53cb3944ec6b4d13a767827f8e87ea1ff739afdaff921f66def5bfcc47a8b9e" Sep 30 15:05:23 crc kubenswrapper[4840]: I0930 15:05:23.972690 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5a51075-4798-48bf-ac82-67a642d1269a-catalog-content\") pod \"e5a51075-4798-48bf-ac82-67a642d1269a\" (UID: \"e5a51075-4798-48bf-ac82-67a642d1269a\") " Sep 30 15:05:23 crc kubenswrapper[4840]: I0930 15:05:23.972904 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5grc\" (UniqueName: \"kubernetes.io/projected/e5a51075-4798-48bf-ac82-67a642d1269a-kube-api-access-t5grc\") pod \"e5a51075-4798-48bf-ac82-67a642d1269a\" (UID: \"e5a51075-4798-48bf-ac82-67a642d1269a\") " Sep 30 15:05:23 crc kubenswrapper[4840]: I0930 15:05:23.973041 4840 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5a51075-4798-48bf-ac82-67a642d1269a-utilities\") pod \"e5a51075-4798-48bf-ac82-67a642d1269a\" (UID: \"e5a51075-4798-48bf-ac82-67a642d1269a\") " Sep 30 15:05:23 crc kubenswrapper[4840]: I0930 15:05:23.974715 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5a51075-4798-48bf-ac82-67a642d1269a-utilities" (OuterVolumeSpecName: "utilities") pod "e5a51075-4798-48bf-ac82-67a642d1269a" (UID: "e5a51075-4798-48bf-ac82-67a642d1269a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:05:23 crc kubenswrapper[4840]: I0930 15:05:23.981003 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5a51075-4798-48bf-ac82-67a642d1269a-kube-api-access-t5grc" (OuterVolumeSpecName: "kube-api-access-t5grc") pod "e5a51075-4798-48bf-ac82-67a642d1269a" (UID: "e5a51075-4798-48bf-ac82-67a642d1269a"). InnerVolumeSpecName "kube-api-access-t5grc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:05:23 crc kubenswrapper[4840]: I0930 15:05:23.998424 4840 scope.go:117] "RemoveContainer" containerID="3210f4d8b25c37367beb5eb6807e2586568932e60d098eaa2175ae7fb13b0a98" Sep 30 15:05:23 crc kubenswrapper[4840]: E0930 15:05:23.999003 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3210f4d8b25c37367beb5eb6807e2586568932e60d098eaa2175ae7fb13b0a98\": container with ID starting with 3210f4d8b25c37367beb5eb6807e2586568932e60d098eaa2175ae7fb13b0a98 not found: ID does not exist" containerID="3210f4d8b25c37367beb5eb6807e2586568932e60d098eaa2175ae7fb13b0a98" Sep 30 15:05:23 crc kubenswrapper[4840]: I0930 15:05:23.999051 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3210f4d8b25c37367beb5eb6807e2586568932e60d098eaa2175ae7fb13b0a98"} err="failed to get container status \"3210f4d8b25c37367beb5eb6807e2586568932e60d098eaa2175ae7fb13b0a98\": rpc error: code = NotFound desc = could not find container \"3210f4d8b25c37367beb5eb6807e2586568932e60d098eaa2175ae7fb13b0a98\": container with ID starting with 3210f4d8b25c37367beb5eb6807e2586568932e60d098eaa2175ae7fb13b0a98 not found: ID does not exist" Sep 30 15:05:23 crc kubenswrapper[4840]: I0930 15:05:23.999078 4840 scope.go:117] "RemoveContainer" containerID="8f002688f13a0db9631486a4b1602411ec8f5de05c142f1a40741b881204a71c" Sep 30 15:05:23 crc kubenswrapper[4840]: E0930 15:05:23.999424 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f002688f13a0db9631486a4b1602411ec8f5de05c142f1a40741b881204a71c\": container with ID starting with 8f002688f13a0db9631486a4b1602411ec8f5de05c142f1a40741b881204a71c not found: ID does not exist" containerID="8f002688f13a0db9631486a4b1602411ec8f5de05c142f1a40741b881204a71c" Sep 30 15:05:23 crc kubenswrapper[4840]: I0930 15:05:23.999473 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f002688f13a0db9631486a4b1602411ec8f5de05c142f1a40741b881204a71c"} err="failed to get container status \"8f002688f13a0db9631486a4b1602411ec8f5de05c142f1a40741b881204a71c\": rpc error: code = NotFound desc = could not find container \"8f002688f13a0db9631486a4b1602411ec8f5de05c142f1a40741b881204a71c\": container with ID starting with 8f002688f13a0db9631486a4b1602411ec8f5de05c142f1a40741b881204a71c not found: ID does not exist" Sep 30 15:05:23 crc kubenswrapper[4840]: I0930 15:05:23.999501 4840 scope.go:117] "RemoveContainer" containerID="c53cb3944ec6b4d13a767827f8e87ea1ff739afdaff921f66def5bfcc47a8b9e" Sep 30 15:05:23 crc kubenswrapper[4840]: E0930 15:05:23.999821 4840 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c53cb3944ec6b4d13a767827f8e87ea1ff739afdaff921f66def5bfcc47a8b9e\": container with ID starting with c53cb3944ec6b4d13a767827f8e87ea1ff739afdaff921f66def5bfcc47a8b9e not found: ID does not exist" containerID="c53cb3944ec6b4d13a767827f8e87ea1ff739afdaff921f66def5bfcc47a8b9e" Sep 30 15:05:23 crc kubenswrapper[4840]: I0930 15:05:23.999842 4840 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c53cb3944ec6b4d13a767827f8e87ea1ff739afdaff921f66def5bfcc47a8b9e"} err="failed to get container status \"c53cb3944ec6b4d13a767827f8e87ea1ff739afdaff921f66def5bfcc47a8b9e\": rpc error: code = NotFound desc = could not find container \"c53cb3944ec6b4d13a767827f8e87ea1ff739afdaff921f66def5bfcc47a8b9e\": container with ID starting with c53cb3944ec6b4d13a767827f8e87ea1ff739afdaff921f66def5bfcc47a8b9e not found: ID does not exist" Sep 30 15:05:24 crc kubenswrapper[4840]: I0930 15:05:24.034142 4840 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5a51075-4798-48bf-ac82-67a642d1269a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e5a51075-4798-48bf-ac82-67a642d1269a" (UID: "e5a51075-4798-48bf-ac82-67a642d1269a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:05:24 crc kubenswrapper[4840]: I0930 15:05:24.075182 4840 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5grc\" (UniqueName: \"kubernetes.io/projected/e5a51075-4798-48bf-ac82-67a642d1269a-kube-api-access-t5grc\") on node \"crc\" DevicePath \"\"" Sep 30 15:05:24 crc kubenswrapper[4840]: I0930 15:05:24.075216 4840 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5a51075-4798-48bf-ac82-67a642d1269a-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 15:05:24 crc kubenswrapper[4840]: I0930 15:05:24.075228 4840 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5a51075-4798-48bf-ac82-67a642d1269a-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 15:05:24 crc kubenswrapper[4840]: I0930 15:05:24.214280 4840 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t2tp4"] Sep 30 15:05:24 crc kubenswrapper[4840]: I0930 15:05:24.226669 4840 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-t2tp4"] Sep 30 15:05:26 crc kubenswrapper[4840]: I0930 15:05:26.116515 4840 scope.go:117] "RemoveContainer" containerID="409430684b16cc7e5d86eb8dbc5c06065665c2c71a8402382eb805da2eca622c" Sep 30 15:05:26 crc kubenswrapper[4840]: I0930 15:05:26.129264 4840 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5a51075-4798-48bf-ac82-67a642d1269a" path="/var/lib/kubelet/pods/e5a51075-4798-48bf-ac82-67a642d1269a/volumes" Sep 30 15:05:26 crc kubenswrapper[4840]: I0930 15:05:26.916483 4840 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-747gk" event={"ID":"10e8b890-7f20-4a36-8e03-898620cf599a","Type":"ContainerStarted","Data":"241c62d14e58d881b513ad29f71b80edabbdbca5ef8fad6b6b22c6e8afe3c302"} var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515066770706024463 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015066770706017400 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015066760236016520 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015066760236015470 5ustar corecore